var/home/core/zuul-output/0000755000175000017500000000000015070106622014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070135151015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006562337615070135141017714 0ustar rootrootOct 04 03:03:06 crc systemd[1]: Starting Kubernetes Kubelet... Oct 04 03:03:07 crc restorecon[4747]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:07 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:03:08 crc restorecon[4747]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 04 03:03:09 crc kubenswrapper[4770]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:03:09 crc kubenswrapper[4770]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 04 03:03:09 crc kubenswrapper[4770]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:03:09 crc kubenswrapper[4770]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:03:09 crc kubenswrapper[4770]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 04 03:03:09 crc kubenswrapper[4770]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.366555 4770 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.370922 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.370949 4770 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.370959 4770 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.370966 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.370973 4770 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.370980 4770 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.370986 4770 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.370993 4770 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371000 4770 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371010 4770 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371035 4770 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371043 4770 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371064 4770 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371072 4770 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371079 4770 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371087 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371094 4770 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371100 4770 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371107 4770 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371114 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371121 4770 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371127 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371132 4770 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371138 4770 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371143 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371148 4770 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371154 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371160 4770 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371166 4770 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371173 4770 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371180 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371185 4770 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371191 4770 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371198 4770 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371205 4770 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371211 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371217 4770 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371225 4770 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371231 4770 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371238 4770 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371243 4770 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371248 4770 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371254 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371260 4770 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371266 4770 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371271 4770 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371276 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371281 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371286 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371292 4770 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371297 4770 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371302 4770 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371307 4770 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371312 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371317 4770 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371323 4770 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371329 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371334 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371339 4770 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371345 4770 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371350 4770 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371355 4770 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371360 4770 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371365 4770 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371370 4770 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371376 4770 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371381 4770 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371386 4770 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371391 4770 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371396 4770 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.371402 4770 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371528 4770 flags.go:64] FLAG: --address="0.0.0.0" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371542 4770 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371553 4770 flags.go:64] FLAG: --anonymous-auth="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371561 4770 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371570 4770 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371580 4770 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371589 4770 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371596 4770 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371603 4770 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371609 4770 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371616 4770 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371623 4770 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371629 4770 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371636 4770 flags.go:64] FLAG: --cgroup-root="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371643 4770 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371650 4770 flags.go:64] FLAG: --client-ca-file="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371656 4770 flags.go:64] FLAG: --cloud-config="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371663 4770 flags.go:64] FLAG: --cloud-provider="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371670 4770 flags.go:64] FLAG: --cluster-dns="[]" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371680 4770 flags.go:64] FLAG: --cluster-domain="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371687 4770 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371693 4770 flags.go:64] FLAG: --config-dir="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371700 4770 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371707 4770 flags.go:64] FLAG: --container-log-max-files="5" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371715 4770 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371723 4770 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371730 4770 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371737 4770 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371743 4770 flags.go:64] FLAG: --contention-profiling="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371749 4770 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371755 4770 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371762 4770 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371769 4770 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371777 4770 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371784 4770 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371791 4770 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371799 4770 flags.go:64] FLAG: --enable-load-reader="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371809 4770 flags.go:64] FLAG: --enable-server="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371817 4770 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371828 4770 flags.go:64] FLAG: --event-burst="100" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371836 4770 flags.go:64] FLAG: --event-qps="50" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371844 4770 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371851 4770 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371859 4770 flags.go:64] FLAG: --eviction-hard="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371869 4770 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371877 4770 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371884 4770 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371892 4770 flags.go:64] FLAG: --eviction-soft="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371899 4770 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371907 4770 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371914 4770 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371920 4770 flags.go:64] FLAG: --experimental-mounter-path="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371927 4770 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371932 4770 flags.go:64] FLAG: --fail-swap-on="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371938 4770 flags.go:64] FLAG: --feature-gates="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371946 4770 flags.go:64] FLAG: --file-check-frequency="20s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371953 4770 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371959 4770 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371965 4770 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371972 4770 flags.go:64] FLAG: --healthz-port="10248" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371978 4770 flags.go:64] FLAG: --help="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371984 4770 flags.go:64] FLAG: --hostname-override="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371991 4770 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.371997 4770 flags.go:64] FLAG: --http-check-frequency="20s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372008 4770 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372033 4770 flags.go:64] FLAG: --image-credential-provider-config="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372040 4770 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372047 4770 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372053 4770 flags.go:64] FLAG: --image-service-endpoint="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372059 4770 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372065 4770 flags.go:64] FLAG: --kube-api-burst="100" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372072 4770 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372078 4770 flags.go:64] FLAG: --kube-api-qps="50" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372085 4770 flags.go:64] FLAG: --kube-reserved="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372091 4770 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372097 4770 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372103 4770 flags.go:64] FLAG: --kubelet-cgroups="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372109 4770 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372116 4770 flags.go:64] FLAG: --lock-file="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372122 4770 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372128 4770 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372134 4770 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372143 4770 flags.go:64] FLAG: --log-json-split-stream="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372150 4770 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372156 4770 flags.go:64] FLAG: --log-text-split-stream="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372162 4770 flags.go:64] FLAG: --logging-format="text" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372167 4770 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372174 4770 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372181 4770 flags.go:64] FLAG: --manifest-url="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372187 4770 flags.go:64] FLAG: --manifest-url-header="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372197 4770 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372203 4770 flags.go:64] FLAG: --max-open-files="1000000" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372211 4770 flags.go:64] FLAG: --max-pods="110" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372217 4770 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372223 4770 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372229 4770 flags.go:64] FLAG: --memory-manager-policy="None" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372236 4770 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372242 4770 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372248 4770 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372255 4770 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372270 4770 flags.go:64] FLAG: --node-status-max-images="50" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372277 4770 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372283 4770 flags.go:64] FLAG: --oom-score-adj="-999" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372290 4770 flags.go:64] FLAG: --pod-cidr="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372297 4770 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372326 4770 flags.go:64] FLAG: --pod-manifest-path="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372332 4770 flags.go:64] FLAG: --pod-max-pids="-1" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372339 4770 flags.go:64] FLAG: --pods-per-core="0" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372346 4770 flags.go:64] FLAG: --port="10250" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372354 4770 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372361 4770 flags.go:64] FLAG: --provider-id="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372368 4770 flags.go:64] FLAG: --qos-reserved="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372375 4770 flags.go:64] FLAG: --read-only-port="10255" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372382 4770 flags.go:64] FLAG: --register-node="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372389 4770 flags.go:64] FLAG: --register-schedulable="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372395 4770 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372406 4770 flags.go:64] FLAG: --registry-burst="10" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372413 4770 flags.go:64] FLAG: --registry-qps="5" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372420 4770 flags.go:64] FLAG: --reserved-cpus="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372427 4770 flags.go:64] FLAG: --reserved-memory="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372435 4770 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372443 4770 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372451 4770 flags.go:64] FLAG: --rotate-certificates="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372458 4770 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372465 4770 flags.go:64] FLAG: --runonce="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372471 4770 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372479 4770 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372486 4770 flags.go:64] FLAG: --seccomp-default="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372492 4770 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372499 4770 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372506 4770 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372512 4770 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372519 4770 flags.go:64] FLAG: --storage-driver-password="root" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372526 4770 flags.go:64] FLAG: --storage-driver-secure="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372533 4770 flags.go:64] FLAG: --storage-driver-table="stats" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372540 4770 flags.go:64] FLAG: --storage-driver-user="root" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372547 4770 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372554 4770 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372561 4770 flags.go:64] FLAG: --system-cgroups="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372567 4770 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372576 4770 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372583 4770 flags.go:64] FLAG: --tls-cert-file="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372589 4770 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372597 4770 flags.go:64] FLAG: --tls-min-version="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372604 4770 flags.go:64] FLAG: --tls-private-key-file="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372611 4770 flags.go:64] FLAG: --topology-manager-policy="none" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372618 4770 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372624 4770 flags.go:64] FLAG: --topology-manager-scope="container" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372630 4770 flags.go:64] FLAG: --v="2" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372638 4770 flags.go:64] FLAG: --version="false" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372647 4770 flags.go:64] FLAG: --vmodule="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372655 4770 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.372661 4770 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372838 4770 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372847 4770 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372854 4770 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372859 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372865 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372873 4770 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372878 4770 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372884 4770 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372889 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372895 4770 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372901 4770 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372906 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372911 4770 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372919 4770 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372924 4770 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372930 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372935 4770 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372941 4770 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372947 4770 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372952 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372958 4770 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372963 4770 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372968 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372974 4770 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372980 4770 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372985 4770 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372991 4770 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.372996 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373037 4770 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373043 4770 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373049 4770 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373056 4770 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373063 4770 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373069 4770 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373075 4770 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373081 4770 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373087 4770 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373092 4770 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373098 4770 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373104 4770 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373117 4770 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373123 4770 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373129 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373134 4770 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373139 4770 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373146 4770 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373152 4770 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373157 4770 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373162 4770 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373168 4770 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373173 4770 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373180 4770 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373187 4770 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373194 4770 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373201 4770 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373209 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373217 4770 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373224 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373232 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373241 4770 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373250 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373257 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373264 4770 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373271 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373280 4770 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373286 4770 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373293 4770 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373299 4770 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373305 4770 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373310 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.373315 4770 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.373335 4770 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.385638 4770 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.385679 4770 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385822 4770 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385837 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385846 4770 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385855 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385863 4770 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385874 4770 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385885 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385895 4770 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385904 4770 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385912 4770 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385921 4770 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385929 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385937 4770 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385945 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385953 4770 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385961 4770 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385969 4770 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385976 4770 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385984 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.385995 4770 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386009 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386045 4770 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386056 4770 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386067 4770 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386079 4770 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386088 4770 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386097 4770 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386105 4770 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386114 4770 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386123 4770 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386131 4770 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386142 4770 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386150 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386159 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386167 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386175 4770 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386183 4770 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386190 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386198 4770 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386207 4770 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386215 4770 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386223 4770 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386231 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386239 4770 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386247 4770 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386255 4770 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386262 4770 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386270 4770 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386278 4770 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386286 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386293 4770 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386303 4770 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386311 4770 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386319 4770 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386327 4770 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386336 4770 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386344 4770 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386352 4770 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386360 4770 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386368 4770 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386376 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386384 4770 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386392 4770 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386400 4770 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386408 4770 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386416 4770 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386424 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386434 4770 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386446 4770 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386455 4770 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386464 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.386478 4770 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386700 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386713 4770 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386722 4770 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386733 4770 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386741 4770 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386749 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386757 4770 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386768 4770 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386779 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386789 4770 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386800 4770 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386811 4770 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386821 4770 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386830 4770 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386840 4770 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386847 4770 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386855 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386863 4770 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386872 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386880 4770 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386888 4770 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386896 4770 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386903 4770 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386911 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386920 4770 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386928 4770 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386936 4770 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386944 4770 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386952 4770 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386960 4770 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386967 4770 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386975 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386983 4770 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.386991 4770 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387001 4770 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387037 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387046 4770 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387054 4770 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387062 4770 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387070 4770 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387078 4770 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387087 4770 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387097 4770 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387107 4770 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387117 4770 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387126 4770 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387135 4770 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387143 4770 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387151 4770 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387160 4770 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387168 4770 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387177 4770 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387186 4770 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387195 4770 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387203 4770 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387213 4770 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387222 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387231 4770 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387238 4770 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387246 4770 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387254 4770 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387262 4770 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387270 4770 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387278 4770 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387285 4770 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387302 4770 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387311 4770 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387319 4770 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387327 4770 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387334 4770 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.387342 4770 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.387354 4770 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.388337 4770 server.go:940] "Client rotation is on, will bootstrap in background" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.394527 4770 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.394660 4770 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.396416 4770 server.go:997] "Starting client certificate rotation" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.396458 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.397421 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-10 06:05:34.438838008 +0000 UTC Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.397578 4770 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2355h2m25.041266936s for next certificate rotation Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.435880 4770 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.439164 4770 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.477795 4770 log.go:25] "Validated CRI v1 runtime API" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.531285 4770 log.go:25] "Validated CRI v1 image API" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.533738 4770 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.540974 4770 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-04-02-58-32-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.541059 4770 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.560942 4770 manager.go:217] Machine: {Timestamp:2025-10-04 03:03:09.557259671 +0000 UTC m=+0.849269403 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3e1fbf36-5852-424f-a63e-18ade8ba99b6 BootID:9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:3b:08:d9 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:3b:08:d9 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:63:cc:cc Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:55:45:c4 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:97:1a:91 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:bf:ef:58 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:6e:63:24 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:06:e2:ab:9b:af:ed Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:1a:79:70:84:04:16 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.561471 4770 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.561754 4770 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.564725 4770 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.565124 4770 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.565188 4770 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.565683 4770 topology_manager.go:138] "Creating topology manager with none policy" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.565701 4770 container_manager_linux.go:303] "Creating device plugin manager" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.566209 4770 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.566266 4770 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.566655 4770 state_mem.go:36] "Initialized new in-memory state store" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.567387 4770 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.571967 4770 kubelet.go:418] "Attempting to sync node with API server" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.571998 4770 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.572040 4770 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.572059 4770 kubelet.go:324] "Adding apiserver pod source" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.572078 4770 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.576504 4770 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.579977 4770 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.581663 4770 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.581609 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.581719 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.583031 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.583084 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.589952 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590082 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590109 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590123 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590148 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590164 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590215 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590239 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590255 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590271 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590316 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.590331 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.592314 4770 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.593127 4770 server.go:1280] "Started kubelet" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.594569 4770 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.594602 4770 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.595411 4770 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.595745 4770 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 04 03:03:09 crc systemd[1]: Started Kubernetes Kubelet. Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.596530 4770 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.596580 4770 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.596646 4770 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 13:26:05.185585811 +0000 UTC Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.596724 4770 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2458h22m55.588868551s for next certificate rotation Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.597102 4770 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.597132 4770 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.597323 4770 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.597361 4770 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.598092 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.598211 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.599256 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="200ms" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.600071 4770 factory.go:55] Registering systemd factory Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.600123 4770 factory.go:221] Registration of the systemd container factory successfully Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.600571 4770 factory.go:153] Registering CRI-O factory Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.600606 4770 factory.go:221] Registration of the crio container factory successfully Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.600727 4770 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.600804 4770 factory.go:103] Registering Raw factory Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.600849 4770 manager.go:1196] Started watching for new ooms in manager Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.602263 4770 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.68:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b2aaa204d87d7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 03:03:09.593069527 +0000 UTC m=+0.885079269,LastTimestamp:2025-10-04 03:03:09.593069527 +0000 UTC m=+0.885079269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.605256 4770 server.go:460] "Adding debug handlers to kubelet server" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.605470 4770 manager.go:319] Starting recovery of all containers Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626540 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626668 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626689 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626710 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626723 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626747 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626764 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626779 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626801 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626819 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626840 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626854 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.626991 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627028 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627046 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627060 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627078 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627094 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627108 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627121 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627135 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627152 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627169 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627182 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627198 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627211 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627233 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627245 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627266 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627284 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627304 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627328 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627349 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627365 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627379 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627399 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627411 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627426 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627442 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627455 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627470 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627481 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627496 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627507 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627517 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.627532 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629659 4770 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629701 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629720 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629745 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629761 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629778 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629794 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629825 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629842 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629859 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629877 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629891 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629909 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629932 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629978 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.629991 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630006 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630032 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630045 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630062 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630079 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630105 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630120 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630132 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630147 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630161 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630179 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630193 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630203 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630251 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630290 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630317 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630336 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630348 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630361 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630377 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630387 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630401 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630412 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630427 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630445 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630546 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630562 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630579 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630591 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630603 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630619 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630632 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630737 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630752 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630765 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630777 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630787 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630800 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630809 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630822 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630836 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630848 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630861 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630882 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630911 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630925 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630943 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630955 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630970 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630986 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.630998 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631031 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631042 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631060 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631080 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631108 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631149 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631164 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631180 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631192 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631206 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631222 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631240 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631251 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631272 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631288 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631318 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631330 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631343 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631355 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631368 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631380 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631393 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631407 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631417 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631431 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631441 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631452 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631468 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631479 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631492 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631501 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631512 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631526 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631536 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631548 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631560 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631572 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631584 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631597 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631611 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631624 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631636 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631652 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631666 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631681 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631692 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631704 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631718 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631730 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631743 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631754 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631771 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631784 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631803 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631816 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631826 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631836 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631850 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631862 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631882 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631893 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631903 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631918 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.631934 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634380 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634498 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634526 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634586 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634611 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634635 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634664 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634692 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634719 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634750 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634784 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634835 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634859 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634896 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634921 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634947 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.634973 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635002 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635067 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635090 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635113 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635144 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635168 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635190 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635213 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635237 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635265 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635302 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635325 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635348 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635373 4770 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635393 4770 reconstruct.go:97] "Volume reconstruction finished" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.635408 4770 reconciler.go:26] "Reconciler: start to sync state" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.650215 4770 manager.go:324] Recovery completed Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.669351 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.669351 4770 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.671485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.671541 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.671559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.672302 4770 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.672365 4770 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.672409 4770 kubelet.go:2335] "Starting kubelet main sync loop" Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.672461 4770 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.672823 4770 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.672840 4770 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.672862 4770 state_mem.go:36] "Initialized new in-memory state store" Oct 04 03:03:09 crc kubenswrapper[4770]: W1004 03:03:09.673072 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.673128 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.698446 4770 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.718742 4770 policy_none.go:49] "None policy: Start" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.720272 4770 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.720296 4770 state_mem.go:35] "Initializing new in-memory state store" Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.772725 4770 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.790199 4770 manager.go:334] "Starting Device Plugin manager" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.790258 4770 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.790272 4770 server.go:79] "Starting device plugin registration server" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.790934 4770 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.790955 4770 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.792465 4770 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.792753 4770 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.792828 4770 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.800784 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="400ms" Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.809378 4770 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.892116 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.893539 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.893608 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.893624 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.893665 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:03:09 crc kubenswrapper[4770]: E1004 03:03:09.894362 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.973164 4770 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.973586 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.976676 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.976718 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.976732 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.976872 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.977108 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.977178 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.977824 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.977880 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.977895 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.978164 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.978179 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.978199 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.978208 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.978320 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.978357 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.979548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.979581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.979655 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.979686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.979710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.979723 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.980044 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.980528 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.980575 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.981353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.981382 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.981395 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.981481 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.981647 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.981664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.981902 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.982050 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.982090 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.984218 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.984250 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.984261 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.984275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.984292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.984275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.984549 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.984586 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.986801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.986870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:09 crc kubenswrapper[4770]: I1004 03:03:09.986885 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041253 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041309 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041347 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041376 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041405 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041430 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041454 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041549 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041588 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041651 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041705 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041828 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041940 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.041999 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.042116 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.095441 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.097496 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.097550 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.097573 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.097619 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:03:10 crc kubenswrapper[4770]: E1004 03:03:10.098432 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.143791 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.143861 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.143891 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.143931 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.143956 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.143981 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144039 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144076 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144078 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144105 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144136 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144156 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144150 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144239 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144170 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144074 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144179 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144178 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144179 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144244 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144331 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144377 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144253 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144405 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144428 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144459 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144466 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144320 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144357 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.144687 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: E1004 03:03:10.202642 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="800ms" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.304994 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.312592 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.331572 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.347175 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.354113 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:10 crc kubenswrapper[4770]: W1004 03:03:10.377085 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4c70f2d3c6032510043cc14e1277834bef236717fb7baff8fa2f42dd5adcac8a WatchSource:0}: Error finding container 4c70f2d3c6032510043cc14e1277834bef236717fb7baff8fa2f42dd5adcac8a: Status 404 returned error can't find the container with id 4c70f2d3c6032510043cc14e1277834bef236717fb7baff8fa2f42dd5adcac8a Oct 04 03:03:10 crc kubenswrapper[4770]: W1004 03:03:10.380305 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4920f9077ce40388ba415a5a0d8fbb7d990d6ff25f11d9b38967b8b55dd2cb79 WatchSource:0}: Error finding container 4920f9077ce40388ba415a5a0d8fbb7d990d6ff25f11d9b38967b8b55dd2cb79: Status 404 returned error can't find the container with id 4920f9077ce40388ba415a5a0d8fbb7d990d6ff25f11d9b38967b8b55dd2cb79 Oct 04 03:03:10 crc kubenswrapper[4770]: W1004 03:03:10.386845 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-5b6265c519c300647275058c8d3cb6038fc7260e06a0e8b2e44e2028d4248881 WatchSource:0}: Error finding container 5b6265c519c300647275058c8d3cb6038fc7260e06a0e8b2e44e2028d4248881: Status 404 returned error can't find the container with id 5b6265c519c300647275058c8d3cb6038fc7260e06a0e8b2e44e2028d4248881 Oct 04 03:03:10 crc kubenswrapper[4770]: W1004 03:03:10.390497 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-9b4dae4fc0256966762f6c2d1d6745d23483ddb1af76cf034f934498ef45bede WatchSource:0}: Error finding container 9b4dae4fc0256966762f6c2d1d6745d23483ddb1af76cf034f934498ef45bede: Status 404 returned error can't find the container with id 9b4dae4fc0256966762f6c2d1d6745d23483ddb1af76cf034f934498ef45bede Oct 04 03:03:10 crc kubenswrapper[4770]: W1004 03:03:10.395173 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-547161c08d4c12eeeb91e53ec526e7fc2d4953591163da404ec484d6b0bd2382 WatchSource:0}: Error finding container 547161c08d4c12eeeb91e53ec526e7fc2d4953591163da404ec484d6b0bd2382: Status 404 returned error can't find the container with id 547161c08d4c12eeeb91e53ec526e7fc2d4953591163da404ec484d6b0bd2382 Oct 04 03:03:10 crc kubenswrapper[4770]: W1004 03:03:10.490577 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:10 crc kubenswrapper[4770]: E1004 03:03:10.490768 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.498668 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.500290 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.500345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.500365 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.500404 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:03:10 crc kubenswrapper[4770]: E1004 03:03:10.501094 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.596539 4770 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:10 crc kubenswrapper[4770]: W1004 03:03:10.638540 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:10 crc kubenswrapper[4770]: E1004 03:03:10.638656 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.679855 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4920f9077ce40388ba415a5a0d8fbb7d990d6ff25f11d9b38967b8b55dd2cb79"} Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.681789 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4c70f2d3c6032510043cc14e1277834bef236717fb7baff8fa2f42dd5adcac8a"} Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.683281 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"547161c08d4c12eeeb91e53ec526e7fc2d4953591163da404ec484d6b0bd2382"} Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.684781 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9b4dae4fc0256966762f6c2d1d6745d23483ddb1af76cf034f934498ef45bede"} Oct 04 03:03:10 crc kubenswrapper[4770]: I1004 03:03:10.688499 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5b6265c519c300647275058c8d3cb6038fc7260e06a0e8b2e44e2028d4248881"} Oct 04 03:03:11 crc kubenswrapper[4770]: E1004 03:03:11.004421 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="1.6s" Oct 04 03:03:11 crc kubenswrapper[4770]: W1004 03:03:11.007089 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:11 crc kubenswrapper[4770]: E1004 03:03:11.007195 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:11 crc kubenswrapper[4770]: W1004 03:03:11.107357 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:11 crc kubenswrapper[4770]: E1004 03:03:11.107501 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.301399 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.303153 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.303195 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.303208 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.303234 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:03:11 crc kubenswrapper[4770]: E1004 03:03:11.303907 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.597191 4770 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.694483 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255"} Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.697186 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8" exitCode=0 Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.697295 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8"} Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.697353 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.699210 4770 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa" exitCode=0 Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.699272 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa"} Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.699394 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.699764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.699805 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.699814 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.700807 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.700828 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.700837 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.701640 4770 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e" exitCode=0 Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.701717 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e"} Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.701735 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.701830 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.702617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.702641 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.702651 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.702931 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.703057 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.703102 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.704555 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c"} Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.704606 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.706438 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.706466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:11 crc kubenswrapper[4770]: I1004 03:03:11.706477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.596929 4770 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:12 crc kubenswrapper[4770]: E1004 03:03:12.605840 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.68:6443: connect: connection refused" interval="3.2s" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.713185 4770 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c" exitCode=0 Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.713256 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.713324 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.713341 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.713354 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.713390 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.714921 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.714966 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.714982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.718491 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.718475 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.718575 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.718590 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.720903 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.720971 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.720996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.726266 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.726326 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.726344 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.729607 4770 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a" exitCode=0 Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.729655 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.729841 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.731353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.731406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.731599 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.732649 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"dc7d55b1cdd86b702936e172a73c1e0f947f6743583a3cb55c02de151a59629a"} Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.732740 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.733799 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.733825 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.733839 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.904964 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.906802 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.906853 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.906866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:12 crc kubenswrapper[4770]: I1004 03:03:12.906903 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:03:12 crc kubenswrapper[4770]: E1004 03:03:12.907314 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.68:6443: connect: connection refused" node="crc" Oct 04 03:03:12 crc kubenswrapper[4770]: W1004 03:03:12.954918 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:12 crc kubenswrapper[4770]: E1004 03:03:12.955166 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:13 crc kubenswrapper[4770]: W1004 03:03:13.361906 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:13 crc kubenswrapper[4770]: E1004 03:03:13.362066 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:13 crc kubenswrapper[4770]: W1004 03:03:13.435192 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:13 crc kubenswrapper[4770]: E1004 03:03:13.435318 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:13 crc kubenswrapper[4770]: E1004 03:03:13.462264 4770 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.68:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b2aaa204d87d7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 03:03:09.593069527 +0000 UTC m=+0.885079269,LastTimestamp:2025-10-04 03:03:09.593069527 +0000 UTC m=+0.885079269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.597314 4770 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.744084 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c7a4395b1e0d72b55a6a110c7492fa5579d39ea7950556e7b59924268d039eb6"} Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.744143 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3"} Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.745523 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.747112 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.747166 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.747184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.748955 4770 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff" exitCode=0 Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.749073 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.749108 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.749141 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.749410 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff"} Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.749471 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.749503 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.751511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.751548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.751585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.751553 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.751603 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.751613 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.751997 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.752062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.752081 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.752278 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.752441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:13 crc kubenswrapper[4770]: I1004 03:03:13.752562 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:13 crc kubenswrapper[4770]: W1004 03:03:13.768922 4770 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.68:6443: connect: connection refused Oct 04 03:03:13 crc kubenswrapper[4770]: E1004 03:03:13.769078 4770 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.68:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.041308 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.111519 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.112093 4770 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.112209 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.760549 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3"} Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.760624 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.760656 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.760707 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.760627 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280"} Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.764340 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.764428 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.764458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.764599 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.764653 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:14 crc kubenswrapper[4770]: I1004 03:03:14.764679 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.772230 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.772267 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860"} Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.772310 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.772333 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.772337 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89"} Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.772482 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f"} Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.774257 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.774325 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.774349 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.774385 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.774454 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.774476 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:15 crc kubenswrapper[4770]: I1004 03:03:15.829869 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.108195 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.109983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.110076 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.110097 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.110133 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.256404 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.775379 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.775521 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.777255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.777324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.777347 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.778448 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.778508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.778533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:16 crc kubenswrapper[4770]: I1004 03:03:16.978640 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.041954 4770 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.042139 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.075483 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.075787 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.077705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.077758 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.077778 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.778557 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.778566 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.780070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.780105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.780121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.780475 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.780515 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:17 crc kubenswrapper[4770]: I1004 03:03:17.780529 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:18 crc kubenswrapper[4770]: I1004 03:03:18.128570 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:18 crc kubenswrapper[4770]: I1004 03:03:18.128788 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:18 crc kubenswrapper[4770]: I1004 03:03:18.130175 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:18 crc kubenswrapper[4770]: I1004 03:03:18.130214 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:18 crc kubenswrapper[4770]: I1004 03:03:18.130229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:19 crc kubenswrapper[4770]: E1004 03:03:19.810420 4770 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 03:03:20 crc kubenswrapper[4770]: I1004 03:03:20.124795 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 04 03:03:20 crc kubenswrapper[4770]: I1004 03:03:20.125085 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:20 crc kubenswrapper[4770]: I1004 03:03:20.126554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:20 crc kubenswrapper[4770]: I1004 03:03:20.126620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:20 crc kubenswrapper[4770]: I1004 03:03:20.126635 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:21 crc kubenswrapper[4770]: I1004 03:03:21.606103 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:21 crc kubenswrapper[4770]: I1004 03:03:21.606402 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:21 crc kubenswrapper[4770]: I1004 03:03:21.608120 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:21 crc kubenswrapper[4770]: I1004 03:03:21.608178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:21 crc kubenswrapper[4770]: I1004 03:03:21.608194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.024704 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.024924 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.026769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.026827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.026846 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.031917 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.793248 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.795121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.795200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.795220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:22 crc kubenswrapper[4770]: I1004 03:03:22.802039 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:23 crc kubenswrapper[4770]: I1004 03:03:23.795221 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:23 crc kubenswrapper[4770]: I1004 03:03:23.796613 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:23 crc kubenswrapper[4770]: I1004 03:03:23.796685 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:23 crc kubenswrapper[4770]: I1004 03:03:23.796706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:24 crc kubenswrapper[4770]: I1004 03:03:24.597122 4770 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 04 03:03:24 crc kubenswrapper[4770]: I1004 03:03:24.801657 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 03:03:24 crc kubenswrapper[4770]: I1004 03:03:24.804820 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c7a4395b1e0d72b55a6a110c7492fa5579d39ea7950556e7b59924268d039eb6" exitCode=255 Oct 04 03:03:24 crc kubenswrapper[4770]: I1004 03:03:24.804882 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c7a4395b1e0d72b55a6a110c7492fa5579d39ea7950556e7b59924268d039eb6"} Oct 04 03:03:24 crc kubenswrapper[4770]: I1004 03:03:24.805185 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:24 crc kubenswrapper[4770]: I1004 03:03:24.806374 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:24 crc kubenswrapper[4770]: I1004 03:03:24.806408 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:24 crc kubenswrapper[4770]: I1004 03:03:24.806419 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:24 crc kubenswrapper[4770]: I1004 03:03:24.806873 4770 scope.go:117] "RemoveContainer" containerID="c7a4395b1e0d72b55a6a110c7492fa5579d39ea7950556e7b59924268d039eb6" Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.038316 4770 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.038385 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.048235 4770 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]log ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]etcd ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/priority-and-fairness-filter ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/start-apiextensions-informers ok Oct 04 03:03:25 crc kubenswrapper[4770]: [-]poststarthook/start-apiextensions-controllers failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [-]poststarthook/crd-informer-synced failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/start-system-namespaces-controller ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 04 03:03:25 crc kubenswrapper[4770]: [-]poststarthook/start-service-ip-repair-controllers failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [-]poststarthook/priority-and-fairness-config-producer failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [-]poststarthook/bootstrap-controller failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/start-kube-aggregator-informers ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 04 03:03:25 crc kubenswrapper[4770]: [-]poststarthook/apiservice-registration-controller failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 04 03:03:25 crc kubenswrapper[4770]: [-]poststarthook/apiservice-discovery-controller failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 04 03:03:25 crc kubenswrapper[4770]: [-]autoregister-completion failed: reason withheld Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/apiservice-openapi-controller ok Oct 04 03:03:25 crc kubenswrapper[4770]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 04 03:03:25 crc kubenswrapper[4770]: livez check failed Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.048308 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.810536 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.813131 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545"} Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.813439 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.814706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.814757 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.814774 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:25 crc kubenswrapper[4770]: I1004 03:03:25.830503 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:26 crc kubenswrapper[4770]: I1004 03:03:26.816761 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:26 crc kubenswrapper[4770]: I1004 03:03:26.817970 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:26 crc kubenswrapper[4770]: I1004 03:03:26.818060 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:26 crc kubenswrapper[4770]: I1004 03:03:26.818077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:27 crc kubenswrapper[4770]: I1004 03:03:27.042344 4770 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 03:03:27 crc kubenswrapper[4770]: I1004 03:03:27.042501 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.117614 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.118495 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.119912 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.120050 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.120135 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.123517 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:29 crc kubenswrapper[4770]: E1004 03:03:29.810857 4770 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.824381 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.825609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.825669 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:29 crc kubenswrapper[4770]: I1004 03:03:29.825687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.033807 4770 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.036271 4770 trace.go:236] Trace[1594360725]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:03:19.325) (total time: 10711ms): Oct 04 03:03:30 crc kubenswrapper[4770]: Trace[1594360725]: ---"Objects listed" error: 10711ms (03:03:30.036) Oct 04 03:03:30 crc kubenswrapper[4770]: Trace[1594360725]: [10.711024966s] [10.711024966s] END Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.036292 4770 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.037223 4770 trace.go:236] Trace[944134916]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:03:19.730) (total time: 10306ms): Oct 04 03:03:30 crc kubenswrapper[4770]: Trace[944134916]: ---"Objects listed" error: 10306ms (03:03:30.037) Oct 04 03:03:30 crc kubenswrapper[4770]: Trace[944134916]: [10.306646539s] [10.306646539s] END Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.037270 4770 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.038771 4770 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.039104 4770 trace.go:236] Trace[726052553]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:03:19.676) (total time: 10362ms): Oct 04 03:03:30 crc kubenswrapper[4770]: Trace[726052553]: ---"Objects listed" error: 10362ms (03:03:30.039) Oct 04 03:03:30 crc kubenswrapper[4770]: Trace[726052553]: [10.362276057s] [10.362276057s] END Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.039124 4770 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.039146 4770 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.039551 4770 trace.go:236] Trace[1532250209]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:03:16.239) (total time: 13799ms): Oct 04 03:03:30 crc kubenswrapper[4770]: Trace[1532250209]: ---"Objects listed" error: 13799ms (03:03:30.039) Oct 04 03:03:30 crc kubenswrapper[4770]: Trace[1532250209]: [13.799868277s] [13.799868277s] END Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.039662 4770 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.185771 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.201881 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.586977 4770 apiserver.go:52] "Watching apiserver" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.591490 4770 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.591904 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.592486 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.592647 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.592716 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.592930 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.593062 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.593096 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.593306 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.593316 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.593450 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.598122 4770 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.601543 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.601733 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.602091 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.602111 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.602202 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.602301 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.603899 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.604102 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.604264 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.638117 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.641825 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.641860 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.641882 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.641904 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.641928 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.642297 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.642320 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.642598 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.642946 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.643033 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.643053 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.643242 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.643411 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.643738 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.643773 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.643808 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.643832 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.644122 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.644216 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.644354 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.644461 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.644509 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.644543 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.644561 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.644817 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.645054 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.644578 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.645139 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.645201 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.645331 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.645360 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.646408 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.646577 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.646700 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.645707 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.645866 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.646858 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647032 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647195 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647313 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647035 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647074 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647254 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647419 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647468 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647515 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647540 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647558 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647575 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647592 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647614 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647631 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647652 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647672 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647691 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647711 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647759 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647777 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647795 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647815 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647828 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647845 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647868 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647921 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647965 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.647988 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648040 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648058 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648076 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648095 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648116 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648136 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648158 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648179 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648197 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648214 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648231 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648249 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648271 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648287 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648304 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648320 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648336 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648353 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648371 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648387 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648404 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648421 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648436 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648453 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648469 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648484 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648501 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648555 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648575 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648595 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648613 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648630 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648646 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648665 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648683 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648700 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650353 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650402 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650429 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650450 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650473 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650598 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650629 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650658 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650688 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650711 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650750 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650777 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650800 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650831 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650858 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650883 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650904 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650930 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650952 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650972 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650996 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651052 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651074 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651101 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651130 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651153 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651213 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651247 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651275 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651304 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651378 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651411 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651441 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651465 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651507 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651535 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651559 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651586 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651624 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651652 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651688 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651770 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651797 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651820 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651849 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651913 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651935 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651969 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651994 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652034 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652058 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652086 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652113 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652134 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652157 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652180 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652199 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652222 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652241 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652265 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652292 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652313 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652336 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652355 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652377 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652399 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652417 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652442 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652464 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652487 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652506 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652575 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652597 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652614 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652677 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652699 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652724 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652744 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652993 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653047 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653070 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653093 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653186 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653208 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653231 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653392 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653413 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653436 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653460 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653511 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653533 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653554 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653576 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653597 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653619 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653646 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653745 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653773 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653796 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653909 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653931 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653958 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654025 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654045 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654091 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654113 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654132 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654188 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654219 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654240 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654261 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654286 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654310 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654330 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654353 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654417 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654448 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654475 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654500 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654520 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654544 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654566 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654591 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654615 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654633 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654659 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654684 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654711 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654729 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654800 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654816 4770 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654826 4770 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654836 4770 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654851 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654862 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654873 4770 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654884 4770 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654898 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654909 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654921 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654931 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654944 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654954 4770 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654964 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654974 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654987 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654998 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.655146 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.655163 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.655200 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.655212 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648178 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648238 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656191 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648508 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656358 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648560 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656432 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650167 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650195 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650313 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650664 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.650796 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651131 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656599 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656620 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656668 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656800 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656843 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651570 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651735 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.651797 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652147 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652447 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652677 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.652773 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653035 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653720 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653642 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.653918 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.654202 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.655844 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.657282 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656127 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656079 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.648437 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656553 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.656976 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.657050 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.657337 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.657246 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.657426 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.657386 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.657596 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.657662 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:03:31.157645445 +0000 UTC m=+22.449655157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.657855 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.657862 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.658075 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.658397 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.658570 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.658612 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.658800 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.658895 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.658978 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.659576 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.659687 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.659726 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.659762 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.660130 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.660221 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.663049 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.663428 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.664061 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.665529 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.666076 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.666089 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.666359 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.666455 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.667054 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.665477 4770 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.667994 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.668087 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.668135 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:31.168109929 +0000 UTC m=+22.460119641 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.668305 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.668503 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.668794 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.668839 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:31.168831438 +0000 UTC m=+22.460841150 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.668859 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.669062 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.669104 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.669160 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.669236 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.669282 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.669305 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.672807 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.673047 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.673222 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.673266 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.673661 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.673830 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.673859 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.674070 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.674306 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.674547 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.674785 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.674999 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.675102 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.675158 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.675163 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.675344 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.667644 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.675657 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.675783 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.676272 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.676608 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.677098 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.677403 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.677642 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.677700 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.677903 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.678719 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.680707 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.680757 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.680801 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.681419 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.682227 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.682640 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.683516 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.684228 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.684333 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.684361 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.684441 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.685104 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.686389 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.686520 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.686762 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.686775 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.686823 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.687760 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.688325 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.688448 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.689821 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.689867 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.690182 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.690185 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.690503 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.690758 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.692266 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.692347 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.692466 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.692561 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.696621 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.696960 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.697000 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.697035 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.697058 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.697115 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:31.197091929 +0000 UTC m=+22.489101861 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.697228 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.697251 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.697263 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.697302 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:31.197292994 +0000 UTC m=+22.489302916 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.697361 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.697649 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.698050 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.698276 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.698385 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.698776 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.698803 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.699399 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.699902 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.699980 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.700569 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.700780 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.703310 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.704308 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.704525 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.704884 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.704985 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.705019 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.705218 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.705628 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.705659 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.706820 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.707256 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.708367 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.708445 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.708592 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.709950 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.710049 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.710124 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.710197 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.711904 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.714841 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.716324 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.716405 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.716408 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.716661 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.716697 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.716853 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.717245 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.717572 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.718101 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.722986 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.723203 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.725585 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.730601 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.737195 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.737643 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.748124 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.755667 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.755824 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.755764 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756222 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756381 4770 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756400 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756414 4770 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756425 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756437 4770 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756446 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756456 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756465 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756475 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756486 4770 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756499 4770 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756526 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756536 4770 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756546 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756558 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756568 4770 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756581 4770 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756592 4770 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756601 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756613 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756623 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756632 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756666 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756694 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756707 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756710 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756752 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756761 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756772 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756780 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756789 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756800 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756808 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756817 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756827 4770 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756836 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756845 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756854 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756864 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756872 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756881 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756892 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756902 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756911 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756920 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756929 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756938 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756949 4770 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756958 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756967 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756977 4770 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756985 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.756994 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757018 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757029 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757038 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757051 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757061 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757070 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757080 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757089 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757098 4770 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757107 4770 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757116 4770 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757125 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757139 4770 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757147 4770 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757156 4770 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757167 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757176 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757187 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757199 4770 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757208 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757218 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757228 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757239 4770 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757248 4770 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757258 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757269 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757281 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757292 4770 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757304 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757324 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757336 4770 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757347 4770 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757360 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757373 4770 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757384 4770 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757400 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757412 4770 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757424 4770 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757437 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757448 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757460 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757474 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757489 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757501 4770 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757512 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757524 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757535 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757547 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757565 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757577 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757590 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757603 4770 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757617 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757630 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757642 4770 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757655 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757668 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757680 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757691 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757703 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757714 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757726 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757740 4770 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757751 4770 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757763 4770 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757774 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757788 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757799 4770 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757811 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757822 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757834 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757846 4770 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757857 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757868 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757879 4770 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757890 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757900 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757913 4770 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757923 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757933 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757943 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757954 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757965 4770 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757976 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.757989 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758000 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758035 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758047 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758059 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758070 4770 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758083 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758095 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758107 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758118 4770 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758135 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758145 4770 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758155 4770 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758165 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758180 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758191 4770 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758205 4770 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758216 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758230 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758242 4770 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758254 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758266 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758277 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758287 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758298 4770 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758311 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758322 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758335 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758345 4770 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758355 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758367 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758377 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758389 4770 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758400 4770 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758411 4770 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.758422 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.768982 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.782284 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.793679 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:30 crc kubenswrapper[4770]: E1004 03:03:30.837321 4770 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.905980 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.914145 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:03:30 crc kubenswrapper[4770]: I1004 03:03:30.920342 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.162000 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.162185 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:03:32.162155776 +0000 UTC m=+23.454165488 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.262940 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.263021 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.263045 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.263070 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263133 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263227 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:32.263205224 +0000 UTC m=+23.555214936 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263246 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263251 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263278 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263297 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263398 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:32.263367468 +0000 UTC m=+23.555377250 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263399 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263447 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:32.263438759 +0000 UTC m=+23.555448581 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263268 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263473 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.263514 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:32.263505301 +0000 UTC m=+23.555515263 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.672745 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.672916 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.679128 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.679898 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.680626 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.681243 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.681769 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.682294 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.682833 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.683375 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.683990 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.684554 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.685102 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.685784 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.686903 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.687647 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.688423 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.689138 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.689804 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.690276 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.692764 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.693726 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.694628 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.695609 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.696507 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.697708 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.698445 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.699552 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.702874 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.703975 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.706155 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.707494 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.707959 4770 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.708073 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.709633 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.710240 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.710642 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.711866 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.712600 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.715069 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.716157 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.717349 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.718037 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.718941 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.719803 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.720702 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.721353 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.722280 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.723081 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.724156 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.724822 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.726497 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.727099 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.727667 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.728257 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.728702 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.832259 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.833146 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.835687 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545" exitCode=255 Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.835808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545"} Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.835906 4770 scope.go:117] "RemoveContainer" containerID="c7a4395b1e0d72b55a6a110c7492fa5579d39ea7950556e7b59924268d039eb6" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.841642 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a670bee2e8a2c02ad1a00cc3f0dadfe919674772f637dc313118b23a7cbb5c20"} Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.848773 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf"} Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.848870 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb"} Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.848888 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"aa6d45e3846013f4432b3fca8593fb1b711029394f09fd7abbfb07cfb3520761"} Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.851536 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6"} Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.851570 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e7eba6b7566e6213687b35f2fffd0f2a836bba66fa0f799c3b41c99545add396"} Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.891388 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.906671 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.929325 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.941163 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.941413 4770 scope.go:117] "RemoveContainer" containerID="90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545" Oct 04 03:03:31 crc kubenswrapper[4770]: E1004 03:03:31.941621 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.947447 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.955948 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.968027 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.979856 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:31 crc kubenswrapper[4770]: I1004 03:03:31.993462 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.004776 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.021954 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.034702 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a4395b1e0d72b55a6a110c7492fa5579d39ea7950556e7b59924268d039eb6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:24Z\\\",\\\"message\\\":\\\"W1004 03:03:13.207426 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 03:03:13.207775 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759546993 cert, and key in /tmp/serving-cert-317156996/serving-signer.crt, /tmp/serving-cert-317156996/serving-signer.key\\\\nI1004 03:03:13.878368 1 observer_polling.go:159] Starting file observer\\\\nW1004 03:03:13.880665 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 03:03:13.880859 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:13.881421 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-317156996/tls.crt::/tmp/serving-cert-317156996/tls.key\\\\\\\"\\\\nF1004 03:03:24.211947 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.048417 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.060929 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.075404 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.087386 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.090789 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.170486 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.170899 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:03:34.17086292 +0000 UTC m=+25.462872632 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.229564 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-xsrdd"] Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.230195 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xsrdd" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.233240 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-h6msx"] Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.233521 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.233603 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-66tn4"] Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.233846 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.233924 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.233595 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.234112 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rtksw"] Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.233765 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.235388 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-26dzh"] Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.236311 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.236516 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.236947 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.236979 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.238586 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.238659 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.238811 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.238969 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.239141 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.239361 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.239718 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.240059 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.240118 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.240168 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.243702 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.244091 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.244260 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.244558 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.244719 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.245253 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.245286 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.269750 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.271670 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.271881 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.271943 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.271989 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.272190 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.272215 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.272233 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.272297 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:34.272276347 +0000 UTC m=+25.564286099 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.272850 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.272873 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.272890 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.272939 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:34.272923683 +0000 UTC m=+25.564933445 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.272995 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.273173 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:34.27314643 +0000 UTC m=+25.565156172 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.273255 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.273295 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:34.273283483 +0000 UTC m=+25.565293235 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.289505 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a4395b1e0d72b55a6a110c7492fa5579d39ea7950556e7b59924268d039eb6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:24Z\\\",\\\"message\\\":\\\"W1004 03:03:13.207426 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 03:03:13.207775 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759546993 cert, and key in /tmp/serving-cert-317156996/serving-signer.crt, /tmp/serving-cert-317156996/serving-signer.key\\\\nI1004 03:03:13.878368 1 observer_polling.go:159] Starting file observer\\\\nW1004 03:03:13.880665 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 03:03:13.880859 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:13.881421 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-317156996/tls.crt::/tmp/serving-cert-317156996/tls.key\\\\\\\"\\\\nF1004 03:03:24.211947 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.306597 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.320237 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.335272 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.347844 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.362126 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373103 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-etc-openvswitch\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373158 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-config\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373185 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg9hp\" (UniqueName: \"kubernetes.io/projected/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-kube-api-access-dg9hp\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373218 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-run-k8s-cni-cncf-io\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373242 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-systemd\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373268 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-ovn-kubernetes\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373292 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hm5p\" (UniqueName: \"kubernetes.io/projected/88d62119-65f3-4846-8b93-53d096fa9df7-kube-api-access-9hm5p\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373319 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovn-node-metrics-cert\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373415 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-hostroot\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373503 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmrx9\" (UniqueName: \"kubernetes.io/projected/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-kube-api-access-pmrx9\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373533 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-ovn\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373553 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-script-lib\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373568 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-kubelet\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373584 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6bd3be93-1791-4cd1-b3ae-b4032548e93a-mcd-auth-proxy-config\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373625 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-os-release\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373684 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-var-lib-kubelet\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373709 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-etc-kubernetes\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373867 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-netns\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373962 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-system-cni-dir\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.373999 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-run-netns\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374157 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/88d62119-65f3-4846-8b93-53d096fa9df7-cni-binary-copy\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374192 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-slash\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374236 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-cnibin\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374303 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-var-lib-cni-multus\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374353 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-openvswitch\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374424 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-bin\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374489 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7a928bfe-7bcc-4d20-9eda-f61f1d0b7733-hosts-file\") pod \"node-resolver-xsrdd\" (UID: \"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\") " pod="openshift-dns/node-resolver-xsrdd" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374515 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-cnibin\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374536 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374554 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-env-overrides\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374572 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-netd\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374590 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374608 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-multus-cni-dir\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374625 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-systemd-units\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374677 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-node-log\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374724 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-log-socket\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374759 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-os-release\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374795 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-multus-conf-dir\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374840 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-system-cni-dir\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374944 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6bd3be93-1791-4cd1-b3ae-b4032548e93a-proxy-tls\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.374994 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-var-lib-cni-bin\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.375078 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-tuning-conf-dir\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.375121 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6bd3be93-1791-4cd1-b3ae-b4032548e93a-rootfs\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.375154 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxsxm\" (UniqueName: \"kubernetes.io/projected/6bd3be93-1791-4cd1-b3ae-b4032548e93a-kube-api-access-gxsxm\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.375188 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-var-lib-openvswitch\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.375421 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-cni-binary-copy\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.375477 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlr9j\" (UniqueName: \"kubernetes.io/projected/7a928bfe-7bcc-4d20-9eda-f61f1d0b7733-kube-api-access-mlr9j\") pod \"node-resolver-xsrdd\" (UID: \"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\") " pod="openshift-dns/node-resolver-xsrdd" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.375513 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-multus-socket-dir-parent\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.375552 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/88d62119-65f3-4846-8b93-53d096fa9df7-multus-daemon-config\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.375585 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-run-multus-certs\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.377064 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.391080 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.403078 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.416257 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.428246 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.444418 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.459278 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476220 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-netd\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476251 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476269 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-multus-cni-dir\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476286 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-systemd-units\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476303 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-node-log\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476321 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-log-socket\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476337 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-os-release\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476356 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-system-cni-dir\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476414 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6bd3be93-1791-4cd1-b3ae-b4032548e93a-proxy-tls\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476432 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-var-lib-cni-bin\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476447 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-multus-conf-dir\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476465 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-tuning-conf-dir\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476482 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6bd3be93-1791-4cd1-b3ae-b4032548e93a-rootfs\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476498 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxsxm\" (UniqueName: \"kubernetes.io/projected/6bd3be93-1791-4cd1-b3ae-b4032548e93a-kube-api-access-gxsxm\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476514 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-var-lib-openvswitch\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476557 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlr9j\" (UniqueName: \"kubernetes.io/projected/7a928bfe-7bcc-4d20-9eda-f61f1d0b7733-kube-api-access-mlr9j\") pod \"node-resolver-xsrdd\" (UID: \"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\") " pod="openshift-dns/node-resolver-xsrdd" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476528 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-netd\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476593 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-log-socket\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476640 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-multus-socket-dir-parent\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476574 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-multus-socket-dir-parent\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476716 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-multus-cni-dir\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476743 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/88d62119-65f3-4846-8b93-53d096fa9df7-multus-daemon-config\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476884 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-run-multus-certs\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476980 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-cni-binary-copy\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477068 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-etc-openvswitch\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477103 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-config\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477136 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg9hp\" (UniqueName: \"kubernetes.io/projected/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-kube-api-access-dg9hp\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477142 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-os-release\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477171 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-run-k8s-cni-cncf-io\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477203 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-systemd\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477218 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-system-cni-dir\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477234 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-ovn-kubernetes\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477265 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovn-node-metrics-cert\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477276 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477302 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-hostroot\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477335 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hm5p\" (UniqueName: \"kubernetes.io/projected/88d62119-65f3-4846-8b93-53d096fa9df7-kube-api-access-9hm5p\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477370 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmrx9\" (UniqueName: \"kubernetes.io/projected/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-kube-api-access-pmrx9\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477409 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-ovn\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477440 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-script-lib\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477468 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-kubelet\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477498 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6bd3be93-1791-4cd1-b3ae-b4032548e93a-mcd-auth-proxy-config\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477529 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-os-release\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477559 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-var-lib-kubelet\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477589 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-etc-kubernetes\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477604 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-run-multus-certs\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477618 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-netns\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477655 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-system-cni-dir\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477683 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-run-netns\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477695 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/88d62119-65f3-4846-8b93-53d096fa9df7-multus-daemon-config\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477748 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-systemd-units\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477773 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/88d62119-65f3-4846-8b93-53d096fa9df7-cni-binary-copy\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477789 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-node-log\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.476671 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-multus-conf-dir\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477807 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-slash\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477855 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-cnibin\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477886 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-var-lib-cni-multus\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477918 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-openvswitch\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477922 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-systemd\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477947 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-bin\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477974 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-ovn-kubernetes\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477982 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7a928bfe-7bcc-4d20-9eda-f61f1d0b7733-hosts-file\") pod \"node-resolver-xsrdd\" (UID: \"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\") " pod="openshift-dns/node-resolver-xsrdd" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478047 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7a928bfe-7bcc-4d20-9eda-f61f1d0b7733-hosts-file\") pod \"node-resolver-xsrdd\" (UID: \"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\") " pod="openshift-dns/node-resolver-xsrdd" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478056 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-cnibin\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478091 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-cnibin\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478095 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-etc-openvswitch\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477858 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-slash\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478105 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478138 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-var-lib-cni-bin\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478144 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-env-overrides\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.477889 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-run-k8s-cni-cncf-io\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478148 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478244 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-kubelet\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478251 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-run-netns\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478267 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-openvswitch\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478318 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-bin\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478498 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-os-release\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478533 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-var-lib-openvswitch\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478552 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-etc-kubernetes\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478069 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-cni-binary-copy\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478588 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6bd3be93-1791-4cd1-b3ae-b4032548e93a-rootfs\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478620 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-cnibin\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478706 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-ovn\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478709 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-var-lib-cni-multus\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478728 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6bd3be93-1791-4cd1-b3ae-b4032548e93a-mcd-auth-proxy-config\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478739 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-config\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478748 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-system-cni-dir\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478780 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-hostroot\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478794 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-netns\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.478808 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/88d62119-65f3-4846-8b93-53d096fa9df7-host-var-lib-kubelet\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.479145 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-env-overrides\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.479326 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-tuning-conf-dir\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.479408 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-script-lib\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.479821 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/88d62119-65f3-4846-8b93-53d096fa9df7-cni-binary-copy\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.482230 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6bd3be93-1791-4cd1-b3ae-b4032548e93a-proxy-tls\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.487500 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovn-node-metrics-cert\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.501497 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmrx9\" (UniqueName: \"kubernetes.io/projected/cc1d26ac-d5d4-4841-a45b-809a9ba074cf-kube-api-access-pmrx9\") pod \"multus-additional-cni-plugins-26dzh\" (UID: \"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\") " pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.505200 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.509093 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlr9j\" (UniqueName: \"kubernetes.io/projected/7a928bfe-7bcc-4d20-9eda-f61f1d0b7733-kube-api-access-mlr9j\") pod \"node-resolver-xsrdd\" (UID: \"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\") " pod="openshift-dns/node-resolver-xsrdd" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.510182 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg9hp\" (UniqueName: \"kubernetes.io/projected/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-kube-api-access-dg9hp\") pod \"ovnkube-node-rtksw\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.510756 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxsxm\" (UniqueName: \"kubernetes.io/projected/6bd3be93-1791-4cd1-b3ae-b4032548e93a-kube-api-access-gxsxm\") pod \"machine-config-daemon-h6msx\" (UID: \"6bd3be93-1791-4cd1-b3ae-b4032548e93a\") " pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.519314 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.520092 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hm5p\" (UniqueName: \"kubernetes.io/projected/88d62119-65f3-4846-8b93-53d096fa9df7-kube-api-access-9hm5p\") pod \"multus-66tn4\" (UID: \"88d62119-65f3-4846-8b93-53d096fa9df7\") " pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.546464 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.558944 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xsrdd" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.563626 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a4395b1e0d72b55a6a110c7492fa5579d39ea7950556e7b59924268d039eb6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:24Z\\\",\\\"message\\\":\\\"W1004 03:03:13.207426 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 03:03:13.207775 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759546993 cert, and key in /tmp/serving-cert-317156996/serving-signer.crt, /tmp/serving-cert-317156996/serving-signer.key\\\\nI1004 03:03:13.878368 1 observer_polling.go:159] Starting file observer\\\\nW1004 03:03:13.880665 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 03:03:13.880859 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:13.881421 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-317156996/tls.crt::/tmp/serving-cert-317156996/tls.key\\\\\\\"\\\\nF1004 03:03:24.211947 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.564654 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.578204 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-66tn4" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.581328 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.591637 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.600973 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-26dzh" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.601225 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.622972 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.644098 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.673564 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.673610 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.673717 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.674128 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:32 crc kubenswrapper[4770]: W1004 03:03:32.678415 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a928bfe_7bcc_4d20_9eda_f61f1d0b7733.slice/crio-938b5324b3ccf7ca54f8dba969d4d0e9700e535f0c542eb0d7374fde04782b5b WatchSource:0}: Error finding container 938b5324b3ccf7ca54f8dba969d4d0e9700e535f0c542eb0d7374fde04782b5b: Status 404 returned error can't find the container with id 938b5324b3ccf7ca54f8dba969d4d0e9700e535f0c542eb0d7374fde04782b5b Oct 04 03:03:32 crc kubenswrapper[4770]: W1004 03:03:32.723497 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc1d26ac_d5d4_4841_a45b_809a9ba074cf.slice/crio-d4e80a6bd0b629b499df848c01f6fc0628bb28494bc53ee585e9706d1a4bd1c3 WatchSource:0}: Error finding container d4e80a6bd0b629b499df848c01f6fc0628bb28494bc53ee585e9706d1a4bd1c3: Status 404 returned error can't find the container with id d4e80a6bd0b629b499df848c01f6fc0628bb28494bc53ee585e9706d1a4bd1c3 Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.857916 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xsrdd" event={"ID":"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733","Type":"ContainerStarted","Data":"938b5324b3ccf7ca54f8dba969d4d0e9700e535f0c542eb0d7374fde04782b5b"} Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.860924 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.863787 4770 scope.go:117] "RemoveContainer" containerID="90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545" Oct 04 03:03:32 crc kubenswrapper[4770]: E1004 03:03:32.863949 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.864627 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" event={"ID":"cc1d26ac-d5d4-4841-a45b-809a9ba074cf","Type":"ContainerStarted","Data":"d4e80a6bd0b629b499df848c01f6fc0628bb28494bc53ee585e9706d1a4bd1c3"} Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.866795 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a"} Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.866841 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"2fd94e16a6226dc24440e4f8e4db5b9cbbd86e0323ba12c7fa7fcbe8c884900b"} Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.868169 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-66tn4" event={"ID":"88d62119-65f3-4846-8b93-53d096fa9df7","Type":"ContainerStarted","Data":"90bcac2f03146f6b81df3f69115ccb46f49179e70b819d42dbc4ba29086cd5c1"} Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.874402 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"01dbf44d602ca1e5e8a26636d9ba1f73601fd9e742f1f6cf59043fb3ee165afb"} Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.878470 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.896693 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.917600 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.936022 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.952563 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.971339 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:32 crc kubenswrapper[4770]: I1004 03:03:32.994616 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:32Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.009719 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.025122 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.049021 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.062800 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.085125 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.101151 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.672763 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:33 crc kubenswrapper[4770]: E1004 03:03:33.672929 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.878688 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff"} Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.880209 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a" exitCode=0 Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.880328 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a"} Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.883558 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-66tn4" event={"ID":"88d62119-65f3-4846-8b93-53d096fa9df7","Type":"ContainerStarted","Data":"96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e"} Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.894456 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7"} Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.894580 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6"} Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.897025 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.897212 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xsrdd" event={"ID":"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733","Type":"ContainerStarted","Data":"a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053"} Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.902853 4770 generic.go:334] "Generic (PLEG): container finished" podID="cc1d26ac-d5d4-4841-a45b-809a9ba074cf" containerID="5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f" exitCode=0 Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.902910 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" event={"ID":"cc1d26ac-d5d4-4841-a45b-809a9ba074cf","Type":"ContainerDied","Data":"5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f"} Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.903588 4770 scope.go:117] "RemoveContainer" containerID="90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545" Oct 04 03:03:33 crc kubenswrapper[4770]: E1004 03:03:33.903742 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.920368 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.939518 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.956333 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.976880 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:33 crc kubenswrapper[4770]: I1004 03:03:33.992610 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:33Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.012217 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.027509 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.042491 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.048049 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.059414 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.060536 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.060933 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.077173 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.101588 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.116516 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.134700 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.155746 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.175398 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.193030 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.197867 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.198073 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:03:38.198056092 +0000 UTC m=+29.490065804 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.206401 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.222609 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.236245 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.251699 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.268180 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.281158 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.294312 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.298886 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.298931 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.298987 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.299033 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299066 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299092 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299105 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299132 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299165 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:38.299147612 +0000 UTC m=+29.591157314 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299164 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299243 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299271 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299285 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299182 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:38.299172832 +0000 UTC m=+29.591182544 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299419 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:38.299351697 +0000 UTC m=+29.591361409 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.299457 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:38.299446669 +0000 UTC m=+29.591456381 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.311381 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.322596 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.335137 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.673528 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.673578 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.674167 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:34 crc kubenswrapper[4770]: E1004 03:03:34.674271 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.911118 4770 generic.go:334] "Generic (PLEG): container finished" podID="cc1d26ac-d5d4-4841-a45b-809a9ba074cf" containerID="4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12" exitCode=0 Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.911254 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" event={"ID":"cc1d26ac-d5d4-4841-a45b-809a9ba074cf","Type":"ContainerDied","Data":"4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12"} Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.916648 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e"} Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.916717 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b"} Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.916742 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d"} Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.916763 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514"} Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.943638 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.965997 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.982226 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:34 crc kubenswrapper[4770]: I1004 03:03:34.997912 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.013641 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.029696 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.045206 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.062029 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.077781 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.096581 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.124555 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.145183 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.178916 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.206036 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.673753 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:35 crc kubenswrapper[4770]: E1004 03:03:35.673957 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.923707 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1"} Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.923751 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce"} Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.925430 4770 generic.go:334] "Generic (PLEG): container finished" podID="cc1d26ac-d5d4-4841-a45b-809a9ba074cf" containerID="619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029" exitCode=0 Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.925455 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" event={"ID":"cc1d26ac-d5d4-4841-a45b-809a9ba074cf","Type":"ContainerDied","Data":"619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029"} Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.952173 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.971945 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:35 crc kubenswrapper[4770]: I1004 03:03:35.997333 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.011543 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.023544 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.037943 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.054029 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.072500 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.090920 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.109383 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.123453 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.139103 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.162826 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.186959 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.439040 4770 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.442952 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.442987 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.442998 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.443133 4770 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.451374 4770 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.451701 4770 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.452895 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.452930 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.452942 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.452960 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.452975 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: E1004 03:03:36.473474 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.478300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.478342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.478357 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.478381 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.478397 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: E1004 03:03:36.491235 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.494693 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.494731 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.494745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.494764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.494776 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: E1004 03:03:36.509119 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.513552 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.513589 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.513602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.513619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.513632 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: E1004 03:03:36.532588 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.536600 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.536669 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.536685 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.536711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.536731 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: E1004 03:03:36.552680 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: E1004 03:03:36.552959 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.555431 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.555527 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.555543 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.555569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.555583 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.659151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.659242 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.659262 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.659288 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.659307 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.673002 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.673103 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:36 crc kubenswrapper[4770]: E1004 03:03:36.673172 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:36 crc kubenswrapper[4770]: E1004 03:03:36.673376 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.765324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.765410 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.765440 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.765470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.765489 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.868404 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.868447 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.868457 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.868475 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.868486 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.933741 4770 generic.go:334] "Generic (PLEG): container finished" podID="cc1d26ac-d5d4-4841-a45b-809a9ba074cf" containerID="098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea" exitCode=0 Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.933827 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" event={"ID":"cc1d26ac-d5d4-4841-a45b-809a9ba074cf","Type":"ContainerDied","Data":"098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea"} Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.953508 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.974697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.974751 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.974767 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.974790 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.974808 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:36Z","lastTransitionTime":"2025-10-04T03:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:36 crc kubenswrapper[4770]: I1004 03:03:36.981491 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.004281 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.019987 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.036064 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.051647 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.067351 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.077972 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.078039 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.078052 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.078070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.078082 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:37Z","lastTransitionTime":"2025-10-04T03:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.080323 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.099297 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.123100 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.136225 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.162716 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.175661 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.180322 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.180350 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.180359 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.180378 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.180390 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:37Z","lastTransitionTime":"2025-10-04T03:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.189335 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.282610 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.282646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.282658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.282676 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.282688 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:37Z","lastTransitionTime":"2025-10-04T03:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.386145 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.386183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.386195 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.386214 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.386228 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:37Z","lastTransitionTime":"2025-10-04T03:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.490053 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.490117 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.490142 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.490174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.490194 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:37Z","lastTransitionTime":"2025-10-04T03:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.593961 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.593995 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.594018 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.594036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.594045 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:37Z","lastTransitionTime":"2025-10-04T03:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.672844 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:37 crc kubenswrapper[4770]: E1004 03:03:37.672985 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.696204 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.696291 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.696313 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.696355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.696374 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:37Z","lastTransitionTime":"2025-10-04T03:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.798814 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.798861 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.798872 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.798892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.798903 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:37Z","lastTransitionTime":"2025-10-04T03:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.901482 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.901517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.901528 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.901546 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.901558 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:37Z","lastTransitionTime":"2025-10-04T03:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.942481 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.946741 4770 generic.go:334] "Generic (PLEG): container finished" podID="cc1d26ac-d5d4-4841-a45b-809a9ba074cf" containerID="64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2" exitCode=0 Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.946782 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" event={"ID":"cc1d26ac-d5d4-4841-a45b-809a9ba074cf","Type":"ContainerDied","Data":"64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2"} Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.966408 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.978114 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:37 crc kubenswrapper[4770]: I1004 03:03:37.997393 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.003707 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.003741 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.003753 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.003770 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.003782 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.012628 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.027038 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.039780 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.048109 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.060309 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.073401 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.089678 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.107081 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.107118 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.107131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.107151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.107162 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.117464 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.131576 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.143805 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.161767 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.209056 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.209099 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.209113 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.209130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.209142 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.247926 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.248148 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:03:46.248133994 +0000 UTC m=+37.540143706 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.312386 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.312484 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.312508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.312544 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.312568 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.349143 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.349402 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.349583 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:46.349546282 +0000 UTC m=+37.641556024 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.349421 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.349780 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.349892 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350103 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350187 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:46.350164648 +0000 UTC m=+37.642174430 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350328 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350425 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350524 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350647 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:46.35062854 +0000 UTC m=+37.642638312 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350818 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350877 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350906 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.350985 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:46.350958729 +0000 UTC m=+37.642968491 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.416423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.416488 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.416507 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.416532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.416548 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.519716 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.519764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.519777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.519797 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.519810 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.623195 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.623249 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.623262 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.623282 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.623298 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.673629 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.673790 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.673625 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:38 crc kubenswrapper[4770]: E1004 03:03:38.674133 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.727047 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.727110 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.727129 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.727158 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.727178 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.833671 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.833880 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.833936 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.834054 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.834295 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.937542 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.937607 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.937630 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.937659 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.937681 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:38Z","lastTransitionTime":"2025-10-04T03:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.962674 4770 generic.go:334] "Generic (PLEG): container finished" podID="cc1d26ac-d5d4-4841-a45b-809a9ba074cf" containerID="9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a" exitCode=0 Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.962763 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" event={"ID":"cc1d26ac-d5d4-4841-a45b-809a9ba074cf","Type":"ContainerDied","Data":"9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a"} Oct 04 03:03:38 crc kubenswrapper[4770]: I1004 03:03:38.984757 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.002755 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.022634 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.031151 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-7wlm6"] Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.031609 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.033969 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.033976 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.034630 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.035256 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.036547 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.040642 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.040671 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.040682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.040700 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.040712 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.052994 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.068721 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.091786 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.111802 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.133982 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.144585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.144668 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.144695 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.144729 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.144751 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.149485 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.158234 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d8687a3-eceb-430c-aed8-ee3bb1493f9d-host\") pod \"node-ca-7wlm6\" (UID: \"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\") " pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.158317 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0d8687a3-eceb-430c-aed8-ee3bb1493f9d-serviceca\") pod \"node-ca-7wlm6\" (UID: \"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\") " pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.158341 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjkwq\" (UniqueName: \"kubernetes.io/projected/0d8687a3-eceb-430c-aed8-ee3bb1493f9d-kube-api-access-zjkwq\") pod \"node-ca-7wlm6\" (UID: \"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\") " pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.162072 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.173942 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.186447 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.200683 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.211969 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.228484 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.240616 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.247618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.247666 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.247685 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.247712 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.247732 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.256056 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.258905 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0d8687a3-eceb-430c-aed8-ee3bb1493f9d-serviceca\") pod \"node-ca-7wlm6\" (UID: \"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\") " pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.258945 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjkwq\" (UniqueName: \"kubernetes.io/projected/0d8687a3-eceb-430c-aed8-ee3bb1493f9d-kube-api-access-zjkwq\") pod \"node-ca-7wlm6\" (UID: \"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\") " pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.258978 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d8687a3-eceb-430c-aed8-ee3bb1493f9d-host\") pod \"node-ca-7wlm6\" (UID: \"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\") " pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.259094 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d8687a3-eceb-430c-aed8-ee3bb1493f9d-host\") pod \"node-ca-7wlm6\" (UID: \"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\") " pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.260081 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0d8687a3-eceb-430c-aed8-ee3bb1493f9d-serviceca\") pod \"node-ca-7wlm6\" (UID: \"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\") " pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.268581 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.285786 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.299367 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjkwq\" (UniqueName: \"kubernetes.io/projected/0d8687a3-eceb-430c-aed8-ee3bb1493f9d-kube-api-access-zjkwq\") pod \"node-ca-7wlm6\" (UID: \"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\") " pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.304808 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.319227 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.332541 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.344231 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.351560 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.351618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.351634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.351653 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.351663 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.355021 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7wlm6" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.357943 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.385860 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.405059 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.425624 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.438170 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.454364 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.454428 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.454444 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.454466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.454478 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.557777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.557823 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.557837 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.557857 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.557873 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.661192 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.661234 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.661249 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.661267 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.661276 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.672973 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:39 crc kubenswrapper[4770]: E1004 03:03:39.673264 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.690895 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.706352 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.728407 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.754369 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.763806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.763855 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.763867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.763884 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.763895 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.793245 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.810329 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.827070 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.839719 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.849642 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.862436 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.866697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.866744 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.866756 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.866777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.866788 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.880195 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.908339 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.924493 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.937665 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.953451 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.967857 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7wlm6" event={"ID":"0d8687a3-eceb-430c-aed8-ee3bb1493f9d","Type":"ContainerStarted","Data":"c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.967926 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7wlm6" event={"ID":"0d8687a3-eceb-430c-aed8-ee3bb1493f9d","Type":"ContainerStarted","Data":"8645856f66cb8214117dcdff6627ffa3af8a97e28eab165ddd1bce2c2e3e8118"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.968318 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.968344 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.968353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.968368 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.968377 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:39Z","lastTransitionTime":"2025-10-04T03:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.972704 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89"} Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.972930 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.972970 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:39 crc kubenswrapper[4770]: I1004 03:03:39.977713 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" event={"ID":"cc1d26ac-d5d4-4841-a45b-809a9ba074cf","Type":"ContainerStarted","Data":"fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.006978 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.021996 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.036252 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.049183 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.051103 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.070223 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.071588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.071633 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.071645 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.071664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.071676 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:40Z","lastTransitionTime":"2025-10-04T03:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.092116 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.111254 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.128643 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.150860 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.165462 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.174675 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.174730 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.174748 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.174775 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.174793 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:40Z","lastTransitionTime":"2025-10-04T03:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.178366 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.192688 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.206415 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.221724 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.251268 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.268155 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.277424 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.277482 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.277502 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.277530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.277549 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:40Z","lastTransitionTime":"2025-10-04T03:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.284944 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.302968 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.317183 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.329517 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.349854 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.366788 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.380903 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.381088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.381449 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.381492 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.381589 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.381617 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:40Z","lastTransitionTime":"2025-10-04T03:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.405889 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.426299 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.447749 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.462512 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.480689 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.484987 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.485050 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.485064 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.485083 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.485096 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:40Z","lastTransitionTime":"2025-10-04T03:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.498022 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.514369 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.588085 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.588153 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.588174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.588205 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.588229 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:40Z","lastTransitionTime":"2025-10-04T03:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.673250 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.673303 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:40 crc kubenswrapper[4770]: E1004 03:03:40.673447 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:40 crc kubenswrapper[4770]: E1004 03:03:40.673556 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.691468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.691523 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.691540 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.691561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.691579 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:40Z","lastTransitionTime":"2025-10-04T03:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.795664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.795733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.795750 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.795782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.795804 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:40Z","lastTransitionTime":"2025-10-04T03:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.901370 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.901468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.901486 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.901511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.901530 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:40Z","lastTransitionTime":"2025-10-04T03:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:40 crc kubenswrapper[4770]: I1004 03:03:40.982289 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.005634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.005706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.005724 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.005750 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.005768 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.022665 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.047169 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.065184 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.087176 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.107686 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.109337 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.109399 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.109416 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.109441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.109460 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.127759 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.148772 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.176210 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.195675 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.219175 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.219266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.219294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.219330 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.219356 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.219895 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.239303 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.257074 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.278624 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.293553 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.313961 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.326645 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.326711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.326726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.326751 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.326766 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.350284 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.430295 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.430347 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.430360 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.430382 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.430397 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.533931 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.533992 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.534010 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.534051 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.534064 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.637252 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.637329 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.637353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.637387 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.637408 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.673091 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:41 crc kubenswrapper[4770]: E1004 03:03:41.673334 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.740836 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.740912 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.740936 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.740970 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.740992 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.844970 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.845035 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.845082 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.845110 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.845128 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.948537 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.948611 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.948636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.948669 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:41 crc kubenswrapper[4770]: I1004 03:03:41.948692 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:41Z","lastTransitionTime":"2025-10-04T03:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.051511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.051569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.051585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.051609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.051621 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.154631 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.154700 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.154719 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.154746 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.154764 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.263090 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.263146 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.263159 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.263190 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.263207 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.366457 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.366900 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.367096 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.367264 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.367422 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.470799 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.470839 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.470849 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.470869 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.470890 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.573295 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.573361 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.573376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.573400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.573415 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.673265 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.673333 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:42 crc kubenswrapper[4770]: E1004 03:03:42.673407 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:42 crc kubenswrapper[4770]: E1004 03:03:42.673512 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.676247 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.676278 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.676286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.676297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.676307 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.779435 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.779498 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.779517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.779543 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.779559 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.883368 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.883468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.883489 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.883518 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.883538 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.986551 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.986594 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.986606 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.986626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:42 crc kubenswrapper[4770]: I1004 03:03:42.986637 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:42Z","lastTransitionTime":"2025-10-04T03:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.090398 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.090483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.090507 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.090534 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.090553 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:43Z","lastTransitionTime":"2025-10-04T03:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.195227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.195261 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.195270 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.195286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.195300 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:43Z","lastTransitionTime":"2025-10-04T03:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.298975 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.299088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.299109 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.299138 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.299156 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:43Z","lastTransitionTime":"2025-10-04T03:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.403047 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.403114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.403136 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.403168 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.403185 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:43Z","lastTransitionTime":"2025-10-04T03:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.506726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.506818 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.506845 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.506882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.506906 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:43Z","lastTransitionTime":"2025-10-04T03:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.610352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.610415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.610432 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.610460 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.610477 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:43Z","lastTransitionTime":"2025-10-04T03:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.673341 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:43 crc kubenswrapper[4770]: E1004 03:03:43.673573 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.713492 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.713563 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.713588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.713619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.713643 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:43Z","lastTransitionTime":"2025-10-04T03:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.817763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.817841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.817867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.817898 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.817921 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:43Z","lastTransitionTime":"2025-10-04T03:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.921327 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.921398 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.921423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.921454 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.921478 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:43Z","lastTransitionTime":"2025-10-04T03:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.995879 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/0.log" Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.999682 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89" exitCode=1 Oct 04 03:03:43 crc kubenswrapper[4770]: I1004 03:03:43.999749 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.000877 4770 scope.go:117] "RemoveContainer" containerID="422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.024321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.024397 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.024423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.024456 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.024475 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.026482 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.049521 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.078086 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.102263 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.126219 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.127698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.127773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.127862 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.127894 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.127914 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.151966 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.171881 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.193234 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.213823 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.232174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.232245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.232263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.232290 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.232306 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.237331 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.275311 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:43Z\\\",\\\"message\\\":\\\"vent handler 5 for removal\\\\nI1004 03:03:43.344486 6041 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:03:43.344528 6041 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:03:43.344537 6041 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:43.344603 6041 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:03:43.344635 6041 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:03:43.344672 6041 factory.go:656] Stopping watch factory\\\\nI1004 03:03:43.344704 6041 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:03:43.344719 6041 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:03:43.344732 6041 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 03:03:43.344743 6041 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:03:43.344755 6041 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:03:43.344766 6041 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:43.344778 6041 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 03:03:43.344942 6041 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.303489 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.332558 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.336873 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.336913 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.336930 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.336955 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.336971 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.355064 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.383100 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.440827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.441146 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.441257 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.441396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.441483 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.545190 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.545286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.545316 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.545351 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.545376 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.648682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.649003 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.649102 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.649196 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.649270 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.673230 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.673238 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:44 crc kubenswrapper[4770]: E1004 03:03:44.673522 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:44 crc kubenswrapper[4770]: E1004 03:03:44.673678 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.753743 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.754090 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.754216 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.754342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.754438 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.858154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.858344 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.858366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.858396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.858416 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.962151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.962226 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.962257 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.962286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:44 crc kubenswrapper[4770]: I1004 03:03:44.962309 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:44Z","lastTransitionTime":"2025-10-04T03:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.007423 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/0.log" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.023344 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.065797 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.065854 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.065872 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.065899 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.065917 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:45Z","lastTransitionTime":"2025-10-04T03:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.124965 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp"] Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.125652 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.129595 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.130569 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.148229 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.169887 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.169954 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.169963 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.169986 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.169998 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:45Z","lastTransitionTime":"2025-10-04T03:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.174667 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.205363 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:43Z\\\",\\\"message\\\":\\\"vent handler 5 for removal\\\\nI1004 03:03:43.344486 6041 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:03:43.344528 6041 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:03:43.344537 6041 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:43.344603 6041 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:03:43.344635 6041 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:03:43.344672 6041 factory.go:656] Stopping watch factory\\\\nI1004 03:03:43.344704 6041 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:03:43.344719 6041 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:03:43.344732 6041 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 03:03:43.344743 6041 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:03:43.344755 6041 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:03:43.344766 6041 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:43.344778 6041 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 03:03:43.344942 6041 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.227996 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.236095 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6f3c536f-6774-43e6-8800-e7e060cd96f2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.236464 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whw6b\" (UniqueName: \"kubernetes.io/projected/6f3c536f-6774-43e6-8800-e7e060cd96f2-kube-api-access-whw6b\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.236697 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6f3c536f-6774-43e6-8800-e7e060cd96f2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.236975 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6f3c536f-6774-43e6-8800-e7e060cd96f2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.248993 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.267915 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.274266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.274327 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.274339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.274362 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.274392 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:45Z","lastTransitionTime":"2025-10-04T03:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.289766 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.314240 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.330191 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.338835 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6f3c536f-6774-43e6-8800-e7e060cd96f2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.338907 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whw6b\" (UniqueName: \"kubernetes.io/projected/6f3c536f-6774-43e6-8800-e7e060cd96f2-kube-api-access-whw6b\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.339022 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6f3c536f-6774-43e6-8800-e7e060cd96f2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.339136 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6f3c536f-6774-43e6-8800-e7e060cd96f2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.340669 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6f3c536f-6774-43e6-8800-e7e060cd96f2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.340820 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6f3c536f-6774-43e6-8800-e7e060cd96f2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.345781 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6f3c536f-6774-43e6-8800-e7e060cd96f2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.347626 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.365943 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whw6b\" (UniqueName: \"kubernetes.io/projected/6f3c536f-6774-43e6-8800-e7e060cd96f2-kube-api-access-whw6b\") pod \"ovnkube-control-plane-749d76644c-z9xqp\" (UID: \"6f3c536f-6774-43e6-8800-e7e060cd96f2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.370297 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.386481 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.386547 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.386562 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.386588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.386605 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:45Z","lastTransitionTime":"2025-10-04T03:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.389116 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.406531 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.419345 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.441805 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.444955 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.460107 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: W1004 03:03:45.467264 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f3c536f_6774_43e6_8800_e7e060cd96f2.slice/crio-82035b18ec16f992339b6a6dbc5c25a13d09e9fe989175752848b7b332e25959 WatchSource:0}: Error finding container 82035b18ec16f992339b6a6dbc5c25a13d09e9fe989175752848b7b332e25959: Status 404 returned error can't find the container with id 82035b18ec16f992339b6a6dbc5c25a13d09e9fe989175752848b7b332e25959 Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.489445 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.489501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.489519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.489548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.489593 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:45Z","lastTransitionTime":"2025-10-04T03:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.592892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.592943 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.592956 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.592973 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.592986 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:45Z","lastTransitionTime":"2025-10-04T03:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.673648 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:45 crc kubenswrapper[4770]: E1004 03:03:45.673923 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.695429 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.695489 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.695510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.695541 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.695562 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:45Z","lastTransitionTime":"2025-10-04T03:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.799789 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.799856 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.799876 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.799905 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.799927 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:45Z","lastTransitionTime":"2025-10-04T03:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.873758 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-krp9c"] Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.874597 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:45 crc kubenswrapper[4770]: E1004 03:03:45.874748 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.895406 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.903147 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.903213 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.903232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.903260 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.903280 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:45Z","lastTransitionTime":"2025-10-04T03:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.924080 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.949243 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.949343 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv2f7\" (UniqueName: \"kubernetes.io/projected/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-kube-api-access-kv2f7\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.960092 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:45 crc kubenswrapper[4770]: I1004 03:03:45.983361 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.003209 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.006303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.006345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.006358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.006377 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.006391 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.024373 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.030882 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" event={"ID":"6f3c536f-6774-43e6-8800-e7e060cd96f2","Type":"ContainerStarted","Data":"4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.030965 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" event={"ID":"6f3c536f-6774-43e6-8800-e7e060cd96f2","Type":"ContainerStarted","Data":"82035b18ec16f992339b6a6dbc5c25a13d09e9fe989175752848b7b332e25959"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.031114 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.051040 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.051119 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv2f7\" (UniqueName: \"kubernetes.io/projected/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-kube-api-access-kv2f7\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.052076 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.052166 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs podName:5e7fe8aa-4b62-47da-8aa7-0d31dcdba457 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:46.55214001 +0000 UTC m=+37.844149912 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs") pod "network-metrics-daemon-krp9c" (UID: "5e7fe8aa-4b62-47da-8aa7-0d31dcdba457") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.055637 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.073622 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv2f7\" (UniqueName: \"kubernetes.io/projected/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-kube-api-access-kv2f7\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.079750 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:43Z\\\",\\\"message\\\":\\\"vent handler 5 for removal\\\\nI1004 03:03:43.344486 6041 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:03:43.344528 6041 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:03:43.344537 6041 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:43.344603 6041 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:03:43.344635 6041 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:03:43.344672 6041 factory.go:656] Stopping watch factory\\\\nI1004 03:03:43.344704 6041 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:03:43.344719 6041 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:03:43.344732 6041 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 03:03:43.344743 6041 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:03:43.344755 6041 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:03:43.344766 6041 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:43.344778 6041 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 03:03:43.344942 6041 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.092812 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.107968 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.109298 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.109330 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.109344 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.109364 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.109377 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.118923 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.154209 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.178124 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.195517 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.212488 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.212542 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.212558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.212579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.212594 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.231429 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.256305 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.256598 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:04:02.256574837 +0000 UTC m=+53.548584539 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.261488 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.291676 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.309968 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.316310 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.316564 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.316627 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.316691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.316758 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.333316 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:43Z\\\",\\\"message\\\":\\\"vent handler 5 for removal\\\\nI1004 03:03:43.344486 6041 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:03:43.344528 6041 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:03:43.344537 6041 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:43.344603 6041 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:03:43.344635 6041 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:03:43.344672 6041 factory.go:656] Stopping watch factory\\\\nI1004 03:03:43.344704 6041 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:03:43.344719 6041 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:03:43.344732 6041 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 03:03:43.344743 6041 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:03:43.344755 6041 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:03:43.344766 6041 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:43.344778 6041 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 03:03:43.344942 6041 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.346905 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.356966 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.357325 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.357369 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.357383 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.357444 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:02.35742686 +0000 UTC m=+53.649436572 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.357289 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.357734 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.357860 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:02.35783897 +0000 UTC m=+53.649848682 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.357751 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.358064 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.358093 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.358264 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.358351 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.358146 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.358482 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:02.358463976 +0000 UTC m=+53.650473688 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.358567 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:02.358554759 +0000 UTC m=+53.650564471 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.369522 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.386821 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.402874 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.419677 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.419728 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.419742 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.419763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.419779 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.437318 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.459255 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.478562 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.496631 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.513275 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.522530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.522569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.522581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.522595 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.522606 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.526382 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.540671 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.552542 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.559921 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.560213 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.560392 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs podName:5e7fe8aa-4b62-47da-8aa7-0d31dcdba457 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:47.560352307 +0000 UTC m=+38.852362219 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs") pod "network-metrics-daemon-krp9c" (UID: "5e7fe8aa-4b62-47da-8aa7-0d31dcdba457") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.570975 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.585076 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.598656 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.626343 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.626423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.626444 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.626475 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.626494 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.673720 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.673803 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.673909 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.674297 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.674721 4770 scope.go:117] "RemoveContainer" containerID="90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.729678 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.729730 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.729743 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.729764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.729779 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.833689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.833763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.833786 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.833818 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.833839 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.849738 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.849773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.849786 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.849804 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.849818 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.870580 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.876912 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.876964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.876973 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.876990 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.877002 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.893289 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.897850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.897884 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.897896 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.897915 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.897927 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.918195 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.924164 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.924240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.924266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.924303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.924343 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.941260 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.945634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.945755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.945820 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.945891 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.945962 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.965730 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:46 crc kubenswrapper[4770]: E1004 03:03:46.965907 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.968865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.968925 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.968940 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.968960 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:46 crc kubenswrapper[4770]: I1004 03:03:46.968974 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:46Z","lastTransitionTime":"2025-10-04T03:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.036840 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/1.log" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.038312 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/0.log" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.047591 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3" exitCode=1 Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.047771 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.047833 4770 scope.go:117] "RemoveContainer" containerID="422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.049657 4770 scope.go:117] "RemoveContainer" containerID="c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3" Oct 04 03:03:47 crc kubenswrapper[4770]: E1004 03:03:47.050011 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.053976 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.057410 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.057840 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.061558 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" event={"ID":"6f3c536f-6774-43e6-8800-e7e060cd96f2","Type":"ContainerStarted","Data":"0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.070694 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.073900 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.073967 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.073995 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.074075 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.074104 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:47Z","lastTransitionTime":"2025-10-04T03:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.097825 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.113562 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.137480 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.151633 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.166479 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.179918 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.179996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.180056 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.180082 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.180099 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:47Z","lastTransitionTime":"2025-10-04T03:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.187111 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.203760 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.217337 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.231683 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.246908 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.268413 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.283402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.283476 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.283492 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.283522 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.283562 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:47Z","lastTransitionTime":"2025-10-04T03:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.287699 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.301121 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.319910 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.346519 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:43Z\\\",\\\"message\\\":\\\"vent handler 5 for removal\\\\nI1004 03:03:43.344486 6041 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:03:43.344528 6041 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:03:43.344537 6041 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:43.344603 6041 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:03:43.344635 6041 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:03:43.344672 6041 factory.go:656] Stopping watch factory\\\\nI1004 03:03:43.344704 6041 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:03:43.344719 6041 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:03:43.344732 6041 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 03:03:43.344743 6041 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:03:43.344755 6041 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:03:43.344766 6041 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:43.344778 6041 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 03:03:43.344942 6041 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:03:46.451263 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.451932 6208 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452745 6208 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452871 6208 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:03:46.453371 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:46.453448 6208 factory.go:656] Stopping watch factory\\\\nI1004 03:03:46.453473 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:46.468711 6208 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:03:46.468757 6208 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:03:46.468860 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:03:46.468890 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:03:46.469006 6208 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.364441 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.386756 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.386801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.386812 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.386829 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.386839 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:47Z","lastTransitionTime":"2025-10-04T03:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.388330 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.407610 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.427269 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.443180 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.478013 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.490120 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.490180 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.490192 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.490212 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.490225 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:47Z","lastTransitionTime":"2025-10-04T03:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.498137 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.514298 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.532844 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.545118 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.559850 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.573596 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.577703 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:47 crc kubenswrapper[4770]: E1004 03:03:47.577910 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:47 crc kubenswrapper[4770]: E1004 03:03:47.578077 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs podName:5e7fe8aa-4b62-47da-8aa7-0d31dcdba457 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:49.578050896 +0000 UTC m=+40.870060598 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs") pod "network-metrics-daemon-krp9c" (UID: "5e7fe8aa-4b62-47da-8aa7-0d31dcdba457") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.592970 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.593020 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.593057 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.593080 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.593093 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:47Z","lastTransitionTime":"2025-10-04T03:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.594991 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.610024 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.624158 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.637559 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.649758 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.669838 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://422c487e927edb5524611894eaebc8ab8003da7673c6f24ca4207fd15f5b3a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:43Z\\\",\\\"message\\\":\\\"vent handler 5 for removal\\\\nI1004 03:03:43.344486 6041 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:03:43.344528 6041 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:03:43.344537 6041 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:43.344603 6041 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:03:43.344635 6041 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:03:43.344672 6041 factory.go:656] Stopping watch factory\\\\nI1004 03:03:43.344704 6041 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:03:43.344719 6041 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:03:43.344732 6041 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 03:03:43.344743 6041 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:03:43.344755 6041 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:03:43.344766 6041 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:43.344778 6041 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 03:03:43.344942 6041 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:03:46.451263 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.451932 6208 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452745 6208 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452871 6208 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:03:46.453371 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:46.453448 6208 factory.go:656] Stopping watch factory\\\\nI1004 03:03:46.453473 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:46.468711 6208 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:03:46.468757 6208 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:03:46.468860 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:03:46.468890 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:03:46.469006 6208 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.673116 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.673181 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:47 crc kubenswrapper[4770]: E1004 03:03:47.673361 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:03:47 crc kubenswrapper[4770]: E1004 03:03:47.673503 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.695720 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.695769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.695792 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.695812 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.695826 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:47Z","lastTransitionTime":"2025-10-04T03:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.798997 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.799127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.799153 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.799191 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.799215 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:47Z","lastTransitionTime":"2025-10-04T03:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.903051 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.903109 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.903123 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.903146 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:47 crc kubenswrapper[4770]: I1004 03:03:47.903159 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:47Z","lastTransitionTime":"2025-10-04T03:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.006433 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.006487 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.006499 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.006517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.006528 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.069057 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/1.log" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.073714 4770 scope.go:117] "RemoveContainer" containerID="c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3" Oct 04 03:03:48 crc kubenswrapper[4770]: E1004 03:03:48.073916 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.093888 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.109923 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.109957 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.109968 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.109982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.109991 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.120177 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:03:46.451263 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.451932 6208 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452745 6208 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452871 6208 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:03:46.453371 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:46.453448 6208 factory.go:656] Stopping watch factory\\\\nI1004 03:03:46.453473 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:46.468711 6208 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:03:46.468757 6208 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:03:46.468860 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:03:46.468890 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:03:46.469006 6208 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.137213 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.158417 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.183376 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.202629 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.213077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.213141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.213164 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.213191 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.213210 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.220978 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.233494 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.249964 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.270167 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.288705 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.305852 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.317339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.317426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.317441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.317468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.317487 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.324970 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.345805 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.407970 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.420384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.420450 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.420467 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.420493 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.420506 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.421514 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.434061 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.524691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.524755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.524771 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.524792 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.524806 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.628423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.628482 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.628500 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.628528 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.628545 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.673057 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.673182 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:48 crc kubenswrapper[4770]: E1004 03:03:48.673242 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:48 crc kubenswrapper[4770]: E1004 03:03:48.673395 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.730939 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.730978 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.730986 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.731000 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.731025 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.833845 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.833904 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.833921 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.833948 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.833966 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.937269 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.937344 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.937368 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.937399 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:48 crc kubenswrapper[4770]: I1004 03:03:48.937425 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:48Z","lastTransitionTime":"2025-10-04T03:03:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.039807 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.039896 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.039933 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.039968 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.039992 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.142942 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.143044 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.143063 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.143090 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.143107 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.245637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.245718 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.245738 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.245765 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.245784 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.348752 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.348831 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.348851 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.348885 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.348903 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.451405 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.451466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.451485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.451516 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.451535 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.554727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.554836 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.554857 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.554949 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.554968 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.611433 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:49 crc kubenswrapper[4770]: E1004 03:03:49.611635 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:49 crc kubenswrapper[4770]: E1004 03:03:49.611767 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs podName:5e7fe8aa-4b62-47da-8aa7-0d31dcdba457 nodeName:}" failed. No retries permitted until 2025-10-04 03:03:53.611739269 +0000 UTC m=+44.903749021 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs") pod "network-metrics-daemon-krp9c" (UID: "5e7fe8aa-4b62-47da-8aa7-0d31dcdba457") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.658397 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.658470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.658488 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.658515 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.658558 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.673149 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.673245 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:49 crc kubenswrapper[4770]: E1004 03:03:49.673313 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:49 crc kubenswrapper[4770]: E1004 03:03:49.673444 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.691784 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.706276 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.723731 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.738312 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.750058 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.767648 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.767710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.767725 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.767747 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.767763 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.791440 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.818138 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.842838 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:03:46.451263 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.451932 6208 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452745 6208 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452871 6208 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:03:46.453371 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:46.453448 6208 factory.go:656] Stopping watch factory\\\\nI1004 03:03:46.453473 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:46.468711 6208 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:03:46.468757 6208 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:03:46.468860 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:03:46.468890 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:03:46.469006 6208 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.853812 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.865564 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.869550 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.869588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.869600 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.869617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.869629 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.877223 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.887399 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.905307 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.918139 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.930630 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.943527 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.958490 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.972924 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.972988 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.973036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.973069 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:49 crc kubenswrapper[4770]: I1004 03:03:49.973092 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:49Z","lastTransitionTime":"2025-10-04T03:03:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.076557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.076612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.076626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.076645 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.076659 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:50Z","lastTransitionTime":"2025-10-04T03:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.181137 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.181197 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.181209 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.181229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.181245 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:50Z","lastTransitionTime":"2025-10-04T03:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.284511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.284563 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.284575 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.284598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.284616 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:50Z","lastTransitionTime":"2025-10-04T03:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.387701 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.387786 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.387810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.387840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.387862 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:50Z","lastTransitionTime":"2025-10-04T03:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.491166 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.491232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.491248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.491273 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.491289 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:50Z","lastTransitionTime":"2025-10-04T03:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.594830 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.594890 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.594905 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.594929 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.594942 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:50Z","lastTransitionTime":"2025-10-04T03:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.673649 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.673677 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:50 crc kubenswrapper[4770]: E1004 03:03:50.673893 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:50 crc kubenswrapper[4770]: E1004 03:03:50.674068 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.697714 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.697782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.697797 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.697821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.697840 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:50Z","lastTransitionTime":"2025-10-04T03:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.801934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.801996 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.802036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.802059 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.802073 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:50Z","lastTransitionTime":"2025-10-04T03:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.905797 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.905894 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.905926 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.905957 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:50 crc kubenswrapper[4770]: I1004 03:03:50.905983 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:50Z","lastTransitionTime":"2025-10-04T03:03:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.009235 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.009296 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.009315 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.009340 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.009362 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.111971 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.112070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.112097 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.112129 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.112159 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.215582 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.215646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.215662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.215689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.215708 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.319137 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.319207 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.319224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.319251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.319269 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.423386 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.423452 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.423471 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.423500 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.423522 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.527154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.527212 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.527230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.527254 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.527270 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.631825 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.632221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.632400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.632527 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.632690 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.673543 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:51 crc kubenswrapper[4770]: E1004 03:03:51.673775 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.674282 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:51 crc kubenswrapper[4770]: E1004 03:03:51.674511 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.735634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.736620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.736986 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.737351 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.737726 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.841060 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.841467 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.841686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.841866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.841994 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.946172 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.946233 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.946252 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.946278 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:51 crc kubenswrapper[4770]: I1004 03:03:51.946296 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:51Z","lastTransitionTime":"2025-10-04T03:03:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.048827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.048892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.048910 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.048937 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.048958 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.151834 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.152318 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.152473 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.152658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.152792 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.256392 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.256474 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.256496 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.256526 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.256550 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.359574 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.359619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.359631 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.359653 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.359664 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.462791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.462852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.462868 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.462953 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.463055 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.566623 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.566739 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.566816 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.566850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.566875 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.671480 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.672560 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.672617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.672656 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.672661 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.672682 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: E1004 03:03:52.672872 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.672932 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:52 crc kubenswrapper[4770]: E1004 03:03:52.673180 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.776513 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.776583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.776608 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.776639 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.776662 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.880200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.880276 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.880301 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.880330 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.880354 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.984107 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.984196 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.984223 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.984297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:52 crc kubenswrapper[4770]: I1004 03:03:52.984317 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:52Z","lastTransitionTime":"2025-10-04T03:03:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.087877 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.088601 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.088664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.088703 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.088728 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:53Z","lastTransitionTime":"2025-10-04T03:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.192287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.192364 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.192379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.192398 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.192409 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:53Z","lastTransitionTime":"2025-10-04T03:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.295928 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.295998 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.296068 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.296097 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.296118 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:53Z","lastTransitionTime":"2025-10-04T03:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.400251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.400324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.400339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.400362 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.400378 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:53Z","lastTransitionTime":"2025-10-04T03:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.503283 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.503337 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.503351 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.503372 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.503385 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:53Z","lastTransitionTime":"2025-10-04T03:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.606065 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.606136 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.606155 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.606181 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.606198 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:53Z","lastTransitionTime":"2025-10-04T03:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.663488 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:53 crc kubenswrapper[4770]: E1004 03:03:53.663782 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:53 crc kubenswrapper[4770]: E1004 03:03:53.663955 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs podName:5e7fe8aa-4b62-47da-8aa7-0d31dcdba457 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:01.663896667 +0000 UTC m=+52.955906539 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs") pod "network-metrics-daemon-krp9c" (UID: "5e7fe8aa-4b62-47da-8aa7-0d31dcdba457") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.672945 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.673064 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:53 crc kubenswrapper[4770]: E1004 03:03:53.673180 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:03:53 crc kubenswrapper[4770]: E1004 03:03:53.673527 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.709674 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.709761 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.709785 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.709815 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.709838 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:53Z","lastTransitionTime":"2025-10-04T03:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.813283 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.813377 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.813395 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.813419 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.813439 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:53Z","lastTransitionTime":"2025-10-04T03:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.916966 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.917057 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.917075 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.917097 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:53 crc kubenswrapper[4770]: I1004 03:03:53.917110 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:53Z","lastTransitionTime":"2025-10-04T03:03:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.020262 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.020323 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.020337 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.020352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.020362 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.124131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.124194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.124204 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.124228 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.124241 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.228233 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.228316 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.228328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.228356 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.228370 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.331774 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.331844 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.331869 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.331893 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.331912 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.435512 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.435609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.435637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.435678 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.435702 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.539296 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.539352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.539368 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.539386 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.539401 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.643506 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.643586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.643607 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.643641 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.643662 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.673740 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.673855 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:54 crc kubenswrapper[4770]: E1004 03:03:54.674079 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:54 crc kubenswrapper[4770]: E1004 03:03:54.674283 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.747561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.747617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.747652 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.747674 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.747687 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.851285 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.851345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.851360 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.851380 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.851394 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.955631 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.955712 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.955737 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.955768 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:54 crc kubenswrapper[4770]: I1004 03:03:54.955791 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:54Z","lastTransitionTime":"2025-10-04T03:03:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.059492 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.059536 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.059544 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.059561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.059573 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.163451 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.163544 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.163564 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.163593 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.163615 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.267100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.267289 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.267321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.267410 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.267489 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.371182 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.371238 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.371252 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.371275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.371290 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.478675 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.478734 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.478746 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.478771 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.478791 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.582534 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.582585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.582599 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.582620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.582634 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.673916 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.673940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:55 crc kubenswrapper[4770]: E1004 03:03:55.674204 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:55 crc kubenswrapper[4770]: E1004 03:03:55.674305 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.686149 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.686887 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.687261 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.687506 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.687745 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.791698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.791763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.791783 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.791810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.791828 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.896050 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.896661 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.896755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.896861 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.896944 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.999668 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.999733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:55 crc kubenswrapper[4770]: I1004 03:03:55.999747 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:55.999771 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:55.999789 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:55Z","lastTransitionTime":"2025-10-04T03:03:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.102664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.102717 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.102728 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.102745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.102756 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:56Z","lastTransitionTime":"2025-10-04T03:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.206062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.206105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.206116 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.206134 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.206149 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:56Z","lastTransitionTime":"2025-10-04T03:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.309100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.309169 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.309189 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.309217 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.309236 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:56Z","lastTransitionTime":"2025-10-04T03:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.412923 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.413985 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.414197 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.414357 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.414507 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:56Z","lastTransitionTime":"2025-10-04T03:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.517629 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.517697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.517720 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.517751 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.517774 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:56Z","lastTransitionTime":"2025-10-04T03:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.621272 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.621763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.622001 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.622263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.622426 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:56Z","lastTransitionTime":"2025-10-04T03:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.673310 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.673484 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:56 crc kubenswrapper[4770]: E1004 03:03:56.673646 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:56 crc kubenswrapper[4770]: E1004 03:03:56.673754 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.726441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.726500 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.726520 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.726545 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.726562 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:56Z","lastTransitionTime":"2025-10-04T03:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.830281 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.830345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.830366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.830390 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.830410 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:56Z","lastTransitionTime":"2025-10-04T03:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.934429 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.934512 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.934538 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.934574 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:56 crc kubenswrapper[4770]: I1004 03:03:56.934601 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:56Z","lastTransitionTime":"2025-10-04T03:03:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.038576 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.038649 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.038672 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.038704 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.038725 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.085329 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.100454 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.118718 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.142568 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.142653 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.142678 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.142714 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.142740 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.143989 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.155571 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.155636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.155659 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.155691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.155717 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.163902 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: E1004 03:03:57.175637 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.181466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.181537 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.181557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.181589 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.181610 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.188944 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: E1004 03:03:57.206212 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.212739 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.212805 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.212825 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.212855 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.212877 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.217161 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: E1004 03:03:57.236415 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.240489 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.241892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.241951 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.241978 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.242055 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.242087 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.264176 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: E1004 03:03:57.266450 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.272489 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.272555 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.272578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.272617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.272642 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.282556 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: E1004 03:03:57.289320 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: E1004 03:03:57.289580 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.292654 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.292727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.292753 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.292785 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.292811 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.306179 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.345814 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:03:46.451263 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.451932 6208 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452745 6208 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452871 6208 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:03:46.453371 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:46.453448 6208 factory.go:656] Stopping watch factory\\\\nI1004 03:03:46.453473 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:46.468711 6208 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:03:46.468757 6208 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:03:46.468860 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:03:46.468890 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:03:46.469006 6208 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.370112 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.392904 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.395837 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.396131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.396152 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.396184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.396204 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.431778 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.456149 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.479166 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.501183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.501278 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.501298 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.501330 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.501350 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.503389 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.524319 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.605780 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.605836 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.605848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.605867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.605881 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.672744 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:57 crc kubenswrapper[4770]: E1004 03:03:57.672975 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.673078 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:57 crc kubenswrapper[4770]: E1004 03:03:57.673272 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.710108 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.710196 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.710210 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.710231 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.710243 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.813562 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.813624 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.813636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.813658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.813673 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.916684 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.916847 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.916908 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.916983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:57 crc kubenswrapper[4770]: I1004 03:03:57.917095 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:57Z","lastTransitionTime":"2025-10-04T03:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.020853 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.020918 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.020931 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.020950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.020964 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.123712 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.123756 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.123769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.123784 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.123796 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.226304 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.226369 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.226382 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.226412 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.226428 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.329482 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.329542 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.329552 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.329571 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.329582 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.432553 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.432607 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.432618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.432638 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.432653 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.541162 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.541244 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.541266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.541294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.541314 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.645892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.645955 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.645973 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.645999 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.646042 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.673347 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.673444 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:03:58 crc kubenswrapper[4770]: E1004 03:03:58.673556 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:03:58 crc kubenswrapper[4770]: E1004 03:03:58.673788 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.750252 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.750326 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.750344 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.750377 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.750397 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.854762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.854846 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.854870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.854897 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.854916 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.958818 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.958891 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.958909 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.958937 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:58 crc kubenswrapper[4770]: I1004 03:03:58.958954 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:58Z","lastTransitionTime":"2025-10-04T03:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.063425 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.065147 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.065187 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.065222 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.065246 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.168932 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.169051 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.169074 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.169092 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.169136 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.272792 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.272867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.272889 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.272918 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.272938 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.376199 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.376248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.376258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.376279 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.376290 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.479537 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.479586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.479595 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.479611 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.479621 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.583363 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.583426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.583436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.583458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.583471 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.673497 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:03:59 crc kubenswrapper[4770]: E1004 03:03:59.673710 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.673522 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:03:59 crc kubenswrapper[4770]: E1004 03:03:59.674105 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.687655 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.687725 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.687752 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.687784 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.687810 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.696958 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.719737 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.740973 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.760844 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.777363 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.790985 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.791048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.791060 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.791081 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.791093 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.800550 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.832594 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:03:46.451263 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.451932 6208 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452745 6208 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452871 6208 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:03:46.453371 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:46.453448 6208 factory.go:656] Stopping watch factory\\\\nI1004 03:03:46.453473 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:46.468711 6208 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:03:46.468757 6208 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:03:46.468860 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:03:46.468890 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:03:46.469006 6208 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.853104 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.871100 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.895496 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.895622 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.895638 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.895658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.895670 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.902831 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.926976 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.947700 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.966858 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.987825 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:03:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.998241 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.998312 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.998324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.998340 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:03:59 crc kubenswrapper[4770]: I1004 03:03:59.998352 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:03:59Z","lastTransitionTime":"2025-10-04T03:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.002769 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.018200 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.036092 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.055713 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.101510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.101560 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.101578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.101602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.101622 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:00Z","lastTransitionTime":"2025-10-04T03:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.205309 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.205612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.205739 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.205944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.206100 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:00Z","lastTransitionTime":"2025-10-04T03:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.308856 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.308911 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.308927 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.308949 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.308961 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:00Z","lastTransitionTime":"2025-10-04T03:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.411546 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.411610 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.411627 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.411653 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.411673 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:00Z","lastTransitionTime":"2025-10-04T03:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.515132 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.516065 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.516314 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.516551 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.516755 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:00Z","lastTransitionTime":"2025-10-04T03:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.620294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.620578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.620714 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.620897 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.621090 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:00Z","lastTransitionTime":"2025-10-04T03:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.673378 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.673418 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:00 crc kubenswrapper[4770]: E1004 03:04:00.673952 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:00 crc kubenswrapper[4770]: E1004 03:04:00.673853 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.723650 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.723705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.723717 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.723738 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.723751 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:00Z","lastTransitionTime":"2025-10-04T03:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.827093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.827150 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.827162 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.827180 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.827192 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:00Z","lastTransitionTime":"2025-10-04T03:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.929760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.929838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.929866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.929903 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:00 crc kubenswrapper[4770]: I1004 03:04:00.929927 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:00Z","lastTransitionTime":"2025-10-04T03:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.032559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.032620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.032637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.032662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.032681 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.135465 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.135535 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.135554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.135581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.135599 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.239334 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.239385 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.239399 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.239419 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.239431 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.342616 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.342687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.342706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.342733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.342756 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.445711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.445767 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.445775 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.445794 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.445805 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.549044 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.549087 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.549114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.549132 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.549142 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.651847 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.651921 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.651944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.651976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.651997 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.673756 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:01 crc kubenswrapper[4770]: E1004 03:04:01.673962 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.674065 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.674486 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:01 crc kubenswrapper[4770]: E1004 03:04:01.674633 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:04:01 crc kubenswrapper[4770]: E1004 03:04:01.674693 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs podName:5e7fe8aa-4b62-47da-8aa7-0d31dcdba457 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:17.67467027 +0000 UTC m=+68.966679972 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs") pod "network-metrics-daemon-krp9c" (UID: "5e7fe8aa-4b62-47da-8aa7-0d31dcdba457") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:04:01 crc kubenswrapper[4770]: E1004 03:04:01.674921 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.675165 4770 scope.go:117] "RemoveContainer" containerID="c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.754795 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.754849 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.754866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.754891 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.754909 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.858390 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.858763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.858781 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.858821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.858839 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.962574 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.962624 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.962637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.962656 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:01 crc kubenswrapper[4770]: I1004 03:04:01.962669 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:01Z","lastTransitionTime":"2025-10-04T03:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.065762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.065831 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.065851 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.065882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.065903 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.137290 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/1.log" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.140657 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.141476 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.169964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.170043 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.170057 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.170077 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.170093 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.173168 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.211881 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:03:46.451263 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.451932 6208 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452745 6208 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452871 6208 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:03:46.453371 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:46.453448 6208 factory.go:656] Stopping watch factory\\\\nI1004 03:03:46.453473 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:46.468711 6208 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:03:46.468757 6208 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:03:46.468860 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:03:46.468890 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:03:46.469006 6208 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.237240 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.259752 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.273159 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.273191 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.273203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.273220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.273231 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.277558 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.281234 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.281674 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:04:34.281657546 +0000 UTC m=+85.573667268 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.288932 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.308373 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.323712 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.336806 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.351230 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.366664 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.376153 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.376207 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.376221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.376242 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.376255 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.382840 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.382924 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.382979 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.383043 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383051 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383125 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383135 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383149 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383157 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383164 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383170 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383145 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:34.383119354 +0000 UTC m=+85.675129116 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383193 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383207 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:34.383198147 +0000 UTC m=+85.675207859 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383220 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:34.383214647 +0000 UTC m=+85.675224359 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.383272 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:34.383249948 +0000 UTC m=+85.675259700 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.384306 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.397638 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.408134 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.422263 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.433887 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.444608 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.456226 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.479207 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.479248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.479260 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.479275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.479286 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.581982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.582088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.582112 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.582143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.582164 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.673350 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.673353 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.673490 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:02 crc kubenswrapper[4770]: E1004 03:04:02.673660 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.684269 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.684303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.684314 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.684331 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.684343 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.788228 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.788318 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.788342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.788376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.788399 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.892316 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.892366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.892383 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.892409 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.892426 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.995583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.995906 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.996072 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.996275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:02 crc kubenswrapper[4770]: I1004 03:04:02.996425 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:02Z","lastTransitionTime":"2025-10-04T03:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.099344 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.099391 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.099400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.099416 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.099427 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:03Z","lastTransitionTime":"2025-10-04T03:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.147843 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/2.log" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.149273 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/1.log" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.153605 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f" exitCode=1 Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.153671 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.153743 4770 scope.go:117] "RemoveContainer" containerID="c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.155900 4770 scope.go:117] "RemoveContainer" containerID="28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f" Oct 04 03:04:03 crc kubenswrapper[4770]: E1004 03:04:03.156223 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.174877 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.192180 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.205001 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.205088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.205105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.205130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.205148 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:03Z","lastTransitionTime":"2025-10-04T03:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.213208 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.229106 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.246156 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.261678 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.278199 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.308920 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1ea2a2b3279d300fadc848479204322204485e1e5dbf10d5460ad57e154f4b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:03:46Z\\\",\\\"message\\\":\\\"io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:03:46.451263 6208 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.451932 6208 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452745 6208 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:03:46.452871 6208 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:03:46.453371 6208 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:03:46.453448 6208 factory.go:656] Stopping watch factory\\\\nI1004 03:03:46.453473 6208 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:03:46.468711 6208 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:03:46.468757 6208 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:03:46.468860 6208 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:03:46.468890 6208 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:03:46.469006 6208 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.309190 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.309244 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.309263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.309287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.309304 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:03Z","lastTransitionTime":"2025-10-04T03:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.325551 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.362859 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.385472 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.404864 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.413093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.413167 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.413184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.413227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.413249 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:03Z","lastTransitionTime":"2025-10-04T03:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.420645 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.440323 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.458140 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.476139 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.490945 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.505706 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.515593 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.515669 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.515697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.515727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.515751 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:03Z","lastTransitionTime":"2025-10-04T03:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.618850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.618935 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.618953 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.618979 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.619084 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:03Z","lastTransitionTime":"2025-10-04T03:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.672767 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.672794 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:03 crc kubenswrapper[4770]: E1004 03:04:03.672920 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:03 crc kubenswrapper[4770]: E1004 03:04:03.673069 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.722459 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.722492 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.722502 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.722516 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.722524 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:03Z","lastTransitionTime":"2025-10-04T03:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.825568 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.825645 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.825662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.825689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.825708 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:03Z","lastTransitionTime":"2025-10-04T03:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.929122 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.929181 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.929200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.929224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:03 crc kubenswrapper[4770]: I1004 03:04:03.929243 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:03Z","lastTransitionTime":"2025-10-04T03:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.031860 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.031926 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.031947 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.031977 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.031994 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.134364 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.134409 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.134421 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.134439 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.134452 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.159536 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/2.log" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.164703 4770 scope.go:117] "RemoveContainer" containerID="28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f" Oct 04 03:04:04 crc kubenswrapper[4770]: E1004 03:04:04.165500 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.177568 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.195639 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.225774 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.237435 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.237476 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.237488 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.237504 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.237519 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.248502 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.261963 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.273876 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.288845 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.300851 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.320222 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.340775 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.340816 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.340829 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.340844 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.340855 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.341220 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.355713 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.369823 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.383928 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.397614 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.425397 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.439730 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.444417 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.444499 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.444517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.444544 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.444563 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.459680 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.477543 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.547221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.547308 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.547329 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.547458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.547487 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.649921 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.649982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.650039 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.650067 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.650085 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.672840 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.672916 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:04 crc kubenswrapper[4770]: E1004 03:04:04.673103 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:04 crc kubenswrapper[4770]: E1004 03:04:04.673309 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.753554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.753620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.753637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.753666 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.753684 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.856640 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.856703 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.856721 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.856747 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.856767 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.960523 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.960600 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.960612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.960635 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:04 crc kubenswrapper[4770]: I1004 03:04:04.960647 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:04Z","lastTransitionTime":"2025-10-04T03:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.063989 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.064061 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.064069 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.064085 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.064093 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.167293 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.167356 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.167374 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.167401 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.167419 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.271694 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.271783 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.271804 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.271831 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.271850 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.375280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.375322 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.375335 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.375359 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.375374 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.478406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.478461 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.478474 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.478495 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.478508 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.581274 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.581346 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.581370 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.581402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.581420 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.672635 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.672706 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:05 crc kubenswrapper[4770]: E1004 03:04:05.672810 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:05 crc kubenswrapper[4770]: E1004 03:04:05.673076 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.684568 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.684612 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.684628 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.684646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.684662 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.787997 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.788099 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.788117 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.788142 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.788159 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.836808 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.854238 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.870564 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.891120 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.891186 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.891207 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.891235 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.891258 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.892749 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.914731 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.934109 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.950912 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.971486 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.993963 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.994057 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.994075 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.994100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:05 crc kubenswrapper[4770]: I1004 03:04:05.994118 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:05Z","lastTransitionTime":"2025-10-04T03:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.002491 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.031273 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.076108 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.097263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.097306 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.097319 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.097338 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.097349 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:06Z","lastTransitionTime":"2025-10-04T03:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.098839 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.111052 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.136571 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.157057 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.177670 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.199545 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.200549 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.200592 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.200609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.200631 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.200646 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:06Z","lastTransitionTime":"2025-10-04T03:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.214709 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.233526 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.303750 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.303810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.303828 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.303857 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.303876 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:06Z","lastTransitionTime":"2025-10-04T03:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.406772 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.406826 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.406840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.406862 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.406875 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:06Z","lastTransitionTime":"2025-10-04T03:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.509686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.509765 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.509786 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.509812 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.509829 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:06Z","lastTransitionTime":"2025-10-04T03:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.613825 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.613894 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.613915 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.613941 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.613958 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:06Z","lastTransitionTime":"2025-10-04T03:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.672967 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.673057 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:06 crc kubenswrapper[4770]: E1004 03:04:06.673140 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:06 crc kubenswrapper[4770]: E1004 03:04:06.673210 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.716386 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.716446 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.716466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.716494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.716516 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:06Z","lastTransitionTime":"2025-10-04T03:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.820206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.820263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.820281 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.820311 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.820333 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:06Z","lastTransitionTime":"2025-10-04T03:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.924101 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.924187 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.924205 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.924236 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:06 crc kubenswrapper[4770]: I1004 03:04:06.924261 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:06Z","lastTransitionTime":"2025-10-04T03:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.027814 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.027887 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.027899 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.027919 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.027933 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.131570 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.131637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.131665 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.131696 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.131719 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.234981 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.235115 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.235159 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.235194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.235217 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.339619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.340001 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.340178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.340413 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.340574 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.444296 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.444624 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.444778 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.444926 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.445089 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.471303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.471510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.471738 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.472178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.472457 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: E1004 03:04:07.492494 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.497951 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.498168 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.498326 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.498483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.498621 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: E1004 03:04:07.519048 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.523966 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.524048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.524066 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.524089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.524105 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: E1004 03:04:07.542250 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.548358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.548405 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.548423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.548446 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.548463 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: E1004 03:04:07.569364 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.575396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.575462 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.575489 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.575522 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.575547 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: E1004 03:04:07.595421 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:07 crc kubenswrapper[4770]: E1004 03:04:07.595645 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.597897 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.597969 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.597986 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.598034 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.598050 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.672923 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.672997 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:07 crc kubenswrapper[4770]: E1004 03:04:07.673159 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:07 crc kubenswrapper[4770]: E1004 03:04:07.673404 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.701188 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.701245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.701262 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.701287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.701304 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.804759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.804821 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.804839 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.804863 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.804884 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.907788 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.907853 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.907865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.907880 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:07 crc kubenswrapper[4770]: I1004 03:04:07.907888 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:07Z","lastTransitionTime":"2025-10-04T03:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.010941 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.010988 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.010998 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.011036 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.011053 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.114376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.114427 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.114448 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.114475 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.114497 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.218401 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.218459 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.218469 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.218489 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.218501 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.321680 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.321721 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.321734 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.321753 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.321766 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.425448 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.425505 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.425518 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.425537 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.425549 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.529328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.529405 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.529428 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.529455 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.529490 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.633392 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.633436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.633447 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.633465 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.633479 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.673594 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.673659 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:08 crc kubenswrapper[4770]: E1004 03:04:08.673820 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:08 crc kubenswrapper[4770]: E1004 03:04:08.674000 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.736635 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.736726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.736751 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.736778 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.736798 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.839671 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.839769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.840106 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.840132 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.840148 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.943791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.943840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.943856 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.943877 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:08 crc kubenswrapper[4770]: I1004 03:04:08.943894 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:08Z","lastTransitionTime":"2025-10-04T03:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.047119 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.047520 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.047669 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.047896 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.048090 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.151491 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.151531 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.151546 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.151566 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.151583 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.254743 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.254779 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.254791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.254809 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.254821 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.357760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.357831 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.357847 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.357875 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.357892 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.460912 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.460980 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.461093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.461121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.461141 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.564375 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.564450 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.564469 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.564493 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.564510 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.668141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.668615 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.668636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.668687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.668722 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.672730 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.672847 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:09 crc kubenswrapper[4770]: E1004 03:04:09.672875 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:09 crc kubenswrapper[4770]: E1004 03:04:09.673141 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.695443 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.713113 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.748905 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.773681 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.773726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.773738 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.773756 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.773769 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.778399 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.791933 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.804561 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.826155 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.840588 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.866280 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.876546 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.876583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.876595 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.876611 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.876621 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.879542 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.893500 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.904343 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.921573 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.936674 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.951667 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.966247 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.981344 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.981436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.981450 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.981478 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.981494 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:09Z","lastTransitionTime":"2025-10-04T03:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:09 crc kubenswrapper[4770]: I1004 03:04:09.995844 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.010769 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.084801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.084850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.084859 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.084876 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.084888 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:10Z","lastTransitionTime":"2025-10-04T03:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.187746 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.188178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.188314 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.188456 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.188580 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:10Z","lastTransitionTime":"2025-10-04T03:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.292093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.292327 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.292360 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.292394 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.292413 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:10Z","lastTransitionTime":"2025-10-04T03:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.396442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.396511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.396529 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.396555 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.396577 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:10Z","lastTransitionTime":"2025-10-04T03:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.501151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.501213 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.501230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.501256 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.501272 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:10Z","lastTransitionTime":"2025-10-04T03:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.605321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.605365 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.605379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.605395 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.605405 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:10Z","lastTransitionTime":"2025-10-04T03:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.673587 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:10 crc kubenswrapper[4770]: E1004 03:04:10.673784 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.673587 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:10 crc kubenswrapper[4770]: E1004 03:04:10.673976 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.708268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.708342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.708364 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.708394 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.708411 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:10Z","lastTransitionTime":"2025-10-04T03:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.811505 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.811566 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.811588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.811618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.811639 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:10Z","lastTransitionTime":"2025-10-04T03:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.915431 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.915493 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.915512 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.915537 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:10 crc kubenswrapper[4770]: I1004 03:04:10.915554 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:10Z","lastTransitionTime":"2025-10-04T03:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.020136 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.020227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.020251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.020285 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.020310 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.124398 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.124466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.124484 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.124509 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.124527 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.227693 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.227747 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.227765 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.227787 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.227803 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.330959 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.331029 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.331048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.331070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.331083 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.434151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.434224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.434244 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.434270 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.434289 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.537449 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.537532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.537559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.537595 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.537621 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.642685 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.642771 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.642795 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.642823 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.642853 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.673253 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.673256 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:11 crc kubenswrapper[4770]: E1004 03:04:11.673441 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:11 crc kubenswrapper[4770]: E1004 03:04:11.673606 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.747517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.747604 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.747627 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.747656 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.747675 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.851130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.851200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.851221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.851249 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.851273 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.953803 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.953853 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.953866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.953884 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:11 crc kubenswrapper[4770]: I1004 03:04:11.953896 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:11Z","lastTransitionTime":"2025-10-04T03:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.056789 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.056851 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.056871 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.056899 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.056916 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.159502 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.159549 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.159562 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.159577 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.159591 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.262577 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.262655 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.262685 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.262714 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.262736 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.367673 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.367741 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.367762 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.367790 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.367812 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.471940 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.472041 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.472067 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.472100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.472125 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.575902 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.575958 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.575975 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.576000 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.576051 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.673276 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.673285 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:12 crc kubenswrapper[4770]: E1004 03:04:12.673499 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:12 crc kubenswrapper[4770]: E1004 03:04:12.673676 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.679118 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.679190 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.679215 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.679247 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.679265 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.782366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.782425 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.782442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.782467 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.782484 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.885395 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.885446 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.885458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.885476 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.885489 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.988818 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.988876 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.988889 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.988907 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:12 crc kubenswrapper[4770]: I1004 03:04:12.988923 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:12Z","lastTransitionTime":"2025-10-04T03:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.092665 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.092753 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.092779 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.092813 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.092838 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:13Z","lastTransitionTime":"2025-10-04T03:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.196624 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.196866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.196890 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.196922 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.196940 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:13Z","lastTransitionTime":"2025-10-04T03:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.300057 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.300134 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.300157 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.300182 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.300200 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:13Z","lastTransitionTime":"2025-10-04T03:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.408770 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.408840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.408850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.408865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.409619 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:13Z","lastTransitionTime":"2025-10-04T03:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.512664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.512718 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.512735 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.512827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.512861 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:13Z","lastTransitionTime":"2025-10-04T03:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.615771 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.615835 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.615853 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.615879 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.615896 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:13Z","lastTransitionTime":"2025-10-04T03:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.673099 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.673110 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:13 crc kubenswrapper[4770]: E1004 03:04:13.673339 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:13 crc kubenswrapper[4770]: E1004 03:04:13.673460 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.718636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.718681 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.718693 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.718715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.718736 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:13Z","lastTransitionTime":"2025-10-04T03:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.821315 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.821404 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.821415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.821429 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.821438 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:13Z","lastTransitionTime":"2025-10-04T03:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.924625 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.924699 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.924724 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.924753 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:13 crc kubenswrapper[4770]: I1004 03:04:13.924775 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:13Z","lastTransitionTime":"2025-10-04T03:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.027345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.027406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.027431 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.027461 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.027483 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.129760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.129827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.129852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.129879 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.129901 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.233350 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.233391 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.233406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.233432 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.233448 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.336081 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.336143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.336160 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.336186 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.336201 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.439094 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.439197 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.439258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.439282 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.439340 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.542413 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.542470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.542488 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.542514 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.542530 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.645659 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.645696 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.645711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.645733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.645754 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.673564 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.673664 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:14 crc kubenswrapper[4770]: E1004 03:04:14.673747 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:14 crc kubenswrapper[4770]: E1004 03:04:14.673853 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.748283 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.748328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.748369 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.748387 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.748398 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.851205 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.851243 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.851253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.851269 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.851278 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.953499 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.953563 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.953580 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.953605 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:14 crc kubenswrapper[4770]: I1004 03:04:14.953622 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:14Z","lastTransitionTime":"2025-10-04T03:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.056272 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.056340 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.056357 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.056382 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.056399 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.159136 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.159166 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.159174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.159186 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.159196 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.261205 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.261239 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.261247 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.261260 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.261269 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.363199 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.363230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.363240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.363253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.363262 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.465700 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.465765 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.465782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.465807 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.465825 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.568384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.568488 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.568501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.568517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.568529 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.670800 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.670859 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.670873 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.670888 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.670897 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.673256 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.673259 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:15 crc kubenswrapper[4770]: E1004 03:04:15.673494 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:15 crc kubenswrapper[4770]: E1004 03:04:15.673636 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.674331 4770 scope.go:117] "RemoveContainer" containerID="28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f" Oct 04 03:04:15 crc kubenswrapper[4770]: E1004 03:04:15.674602 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.780080 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.780136 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.780153 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.780178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.780196 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.882851 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.882927 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.882950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.882979 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.882999 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.986238 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.986306 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.986330 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.986363 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:15 crc kubenswrapper[4770]: I1004 03:04:15.986386 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:15Z","lastTransitionTime":"2025-10-04T03:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.089418 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.089461 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.089470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.089485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.089495 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:16Z","lastTransitionTime":"2025-10-04T03:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.192212 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.192253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.192264 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.192280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.192292 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:16Z","lastTransitionTime":"2025-10-04T03:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.295048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.295094 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.295104 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.295121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.295131 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:16Z","lastTransitionTime":"2025-10-04T03:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.398663 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.398698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.398707 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.398721 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.398731 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:16Z","lastTransitionTime":"2025-10-04T03:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.501881 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.501961 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.501983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.502048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.502074 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:16Z","lastTransitionTime":"2025-10-04T03:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.605670 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.605752 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.605776 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.605813 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.605838 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:16Z","lastTransitionTime":"2025-10-04T03:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.673356 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:16 crc kubenswrapper[4770]: E1004 03:04:16.673502 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.673618 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:16 crc kubenswrapper[4770]: E1004 03:04:16.673860 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.708495 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.708532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.708541 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.708556 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.708565 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:16Z","lastTransitionTime":"2025-10-04T03:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.811056 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.811119 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.811135 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.811161 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.811181 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:16Z","lastTransitionTime":"2025-10-04T03:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.914584 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.914666 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.914686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.914712 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:16 crc kubenswrapper[4770]: I1004 03:04:16.914736 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:16Z","lastTransitionTime":"2025-10-04T03:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.018100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.018200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.018220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.018246 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.018269 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.121356 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.121483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.121510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.121539 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.121557 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.224697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.224732 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.224742 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.224757 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.224767 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.328159 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.328246 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.328261 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.328286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.328300 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.432439 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.432552 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.432581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.432619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.432645 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.535972 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.536114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.536126 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.536145 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.536158 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.639322 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.639378 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.639392 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.639415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.639428 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.672988 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.673081 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.673150 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.673320 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.698141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.698206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.698220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.698250 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.698267 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.712724 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.717263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.717309 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.717320 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.717341 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.717358 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.731090 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.735663 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.735722 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.735734 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.735753 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.735766 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.752374 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.760666 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.760723 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.760737 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.760759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.760774 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.772651 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.772885 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.773022 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs podName:5e7fe8aa-4b62-47da-8aa7-0d31dcdba457 nodeName:}" failed. No retries permitted until 2025-10-04 03:04:49.772970778 +0000 UTC m=+101.064980640 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs") pod "network-metrics-daemon-krp9c" (UID: "5e7fe8aa-4b62-47da-8aa7-0d31dcdba457") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.773225 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.777094 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.777133 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.777183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.777209 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.777222 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.792219 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:17 crc kubenswrapper[4770]: E1004 03:04:17.792386 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.794174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.794227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.794246 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.794275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.794299 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.898141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.898185 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.898194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.898214 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:17 crc kubenswrapper[4770]: I1004 03:04:17.898226 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:17Z","lastTransitionTime":"2025-10-04T03:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.001485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.001540 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.001555 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.001577 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.001591 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.105494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.105560 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.105581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.105605 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.105625 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.210140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.210219 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.210233 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.210254 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.210265 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.312885 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.312949 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.312961 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.312982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.312994 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.416230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.416305 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.416325 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.416355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.416373 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.519420 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.519472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.519483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.519503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.519516 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.622482 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.622540 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.622553 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.622569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.622579 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.673482 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.673482 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:18 crc kubenswrapper[4770]: E1004 03:04:18.673694 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:18 crc kubenswrapper[4770]: E1004 03:04:18.673819 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.725820 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.725868 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.725888 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.725913 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.725930 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.829300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.829358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.829373 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.829391 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.829405 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.932664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.932727 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.932740 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.932761 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:18 crc kubenswrapper[4770]: I1004 03:04:18.932780 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:18Z","lastTransitionTime":"2025-10-04T03:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.036111 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.036227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.036251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.036281 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.036307 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.139366 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.139402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.139415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.139431 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.139439 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.241472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.241503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.241512 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.241525 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.241534 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.345044 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.345095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.345108 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.345126 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.345140 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.448735 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.448789 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.448801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.448820 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.448834 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.551757 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.551804 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.551816 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.551835 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.551846 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.654355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.654403 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.654413 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.654430 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.654447 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.672745 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:19 crc kubenswrapper[4770]: E1004 03:04:19.672865 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.673047 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:19 crc kubenswrapper[4770]: E1004 03:04:19.673269 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.701199 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.718283 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.734824 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.756704 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.758107 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.758143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.758155 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.758177 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.758641 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.770980 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.784040 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.801660 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.817611 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.834788 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.850801 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.861272 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.861323 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.861335 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.861352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.861363 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.868998 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.884504 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.895816 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.911572 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.926877 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.941861 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.961282 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.964872 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.964935 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.964949 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.964972 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.964990 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:19Z","lastTransitionTime":"2025-10-04T03:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:19 crc kubenswrapper[4770]: I1004 03:04:19.978207 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: E1004 03:04:20.061686 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88d62119_65f3_4846_8b93_53d096fa9df7.slice/crio-conmon-96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.068620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.068681 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.068696 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.068717 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.068727 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:20Z","lastTransitionTime":"2025-10-04T03:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.171658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.171743 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.171755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.171779 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.171796 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:20Z","lastTransitionTime":"2025-10-04T03:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.221638 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/0.log" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.221706 4770 generic.go:334] "Generic (PLEG): container finished" podID="88d62119-65f3-4846-8b93-53d096fa9df7" containerID="96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e" exitCode=1 Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.221748 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-66tn4" event={"ID":"88d62119-65f3-4846-8b93-53d096fa9df7","Type":"ContainerDied","Data":"96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e"} Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.222259 4770 scope.go:117] "RemoveContainer" containerID="96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.237848 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.256057 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:20Z\\\",\\\"message\\\":\\\"2025-10-04T03:03:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3\\\\n2025-10-04T03:03:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3 to /host/opt/cni/bin/\\\\n2025-10-04T03:03:34Z [verbose] multus-daemon started\\\\n2025-10-04T03:03:34Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:04:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.271116 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.277316 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.277367 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.277380 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.277398 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.277412 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:20Z","lastTransitionTime":"2025-10-04T03:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.280594 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.291156 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.300377 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.312478 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.325128 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.336136 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.347721 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.374597 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.381088 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.381130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.381164 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.381189 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.381203 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:20Z","lastTransitionTime":"2025-10-04T03:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.385999 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.399436 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.412227 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.425543 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.440187 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.465269 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.484474 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.484751 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.484770 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.484805 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.484820 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.484831 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:20Z","lastTransitionTime":"2025-10-04T03:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.588358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.588406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.588417 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.588438 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.588449 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:20Z","lastTransitionTime":"2025-10-04T03:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.673419 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.673509 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:20 crc kubenswrapper[4770]: E1004 03:04:20.673768 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:20 crc kubenswrapper[4770]: E1004 03:04:20.673921 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.691848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.691948 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.692160 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.692326 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.692499 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:20Z","lastTransitionTime":"2025-10-04T03:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.795660 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.795941 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.796087 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.796195 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.796273 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:20Z","lastTransitionTime":"2025-10-04T03:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.898734 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.898770 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.898782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.898797 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:20 crc kubenswrapper[4770]: I1004 03:04:20.898807 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:20Z","lastTransitionTime":"2025-10-04T03:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.001459 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.001519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.001533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.001556 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.001572 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.104594 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.104676 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.104696 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.104724 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.104743 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.207579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.207617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.207630 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.207646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.207657 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.225731 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/0.log" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.225783 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-66tn4" event={"ID":"88d62119-65f3-4846-8b93-53d096fa9df7","Type":"ContainerStarted","Data":"6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.247020 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.266177 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.286036 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.304084 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.309839 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.309886 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.309902 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.309925 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.309946 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.322297 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.336935 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.351139 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.363754 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:20Z\\\",\\\"message\\\":\\\"2025-10-04T03:03:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3\\\\n2025-10-04T03:03:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3 to /host/opt/cni/bin/\\\\n2025-10-04T03:03:34Z [verbose] multus-daemon started\\\\n2025-10-04T03:03:34Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:04:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:04:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.377092 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.392413 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.405786 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.411773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.411811 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.411824 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.411841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.411852 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.419559 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.429341 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.441636 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.454221 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.468692 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.489622 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.501786 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.513623 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.513681 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.513694 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.513708 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.513718 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.616132 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.616190 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.616211 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.616240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.616260 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.673303 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.673381 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:21 crc kubenswrapper[4770]: E1004 03:04:21.673457 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:21 crc kubenswrapper[4770]: E1004 03:04:21.673532 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.719341 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.719406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.719426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.719452 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.719471 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.822089 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.822148 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.822167 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.822184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.822193 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.924461 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.924509 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.924521 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.924539 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:21 crc kubenswrapper[4770]: I1004 03:04:21.924551 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:21Z","lastTransitionTime":"2025-10-04T03:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.027396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.027467 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.027479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.027518 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.027532 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.130733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.130768 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.130777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.130792 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.130803 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.233123 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.233208 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.233226 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.233874 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.233903 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.337522 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.337573 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.337591 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.337615 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.337632 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.439658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.439698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.439708 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.439726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.439737 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.543194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.543238 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.543248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.543263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.543274 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.645748 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.645809 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.645827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.645853 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.645871 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.673313 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.673341 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:22 crc kubenswrapper[4770]: E1004 03:04:22.673508 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:22 crc kubenswrapper[4770]: E1004 03:04:22.673604 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.748850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.748904 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.748921 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.748945 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.748965 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.852146 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.852184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.852194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.852208 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.852217 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.955369 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.955446 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.955464 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.955490 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:22 crc kubenswrapper[4770]: I1004 03:04:22.955508 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:22Z","lastTransitionTime":"2025-10-04T03:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.058707 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.058786 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.058810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.058841 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.058863 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:23Z","lastTransitionTime":"2025-10-04T03:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.162083 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.162119 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.162129 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.162145 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.162158 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:23Z","lastTransitionTime":"2025-10-04T03:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.264324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.264381 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.264397 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.264416 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.264429 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:23Z","lastTransitionTime":"2025-10-04T03:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.367456 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.367528 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.367546 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.367570 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.367588 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:23Z","lastTransitionTime":"2025-10-04T03:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.470686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.470756 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.470780 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.470811 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.470830 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:23Z","lastTransitionTime":"2025-10-04T03:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.574600 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.574665 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.574697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.574723 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.574743 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:23Z","lastTransitionTime":"2025-10-04T03:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.677168 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:23 crc kubenswrapper[4770]: E1004 03:04:23.677339 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.677613 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:23 crc kubenswrapper[4770]: E1004 03:04:23.678057 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.680799 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.680860 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.680882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.680914 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.680959 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:23Z","lastTransitionTime":"2025-10-04T03:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.784152 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.784195 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.784204 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.784222 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.784231 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:23Z","lastTransitionTime":"2025-10-04T03:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.887831 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.887879 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.887895 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.887919 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:23 crc kubenswrapper[4770]: I1004 03:04:23.887935 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:23Z","lastTransitionTime":"2025-10-04T03:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.001804 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.001898 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.001919 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.001954 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.001984 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.105946 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.105999 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.106045 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.106072 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.106090 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.210104 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.210226 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.210485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.210510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.210527 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.314095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.314172 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.314191 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.314215 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.314235 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.417221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.417288 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.417311 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.417339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.417363 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.521080 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.521153 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.521175 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.521202 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.521220 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.624105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.624161 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.624178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.624200 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.624217 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.673162 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.673525 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:24 crc kubenswrapper[4770]: E1004 03:04:24.673639 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:24 crc kubenswrapper[4770]: E1004 03:04:24.673967 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.691262 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.728337 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.728687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.728826 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.728956 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.729182 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.832539 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.832602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.832621 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.832645 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.832661 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.935763 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.935828 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.935845 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.935870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:24 crc kubenswrapper[4770]: I1004 03:04:24.935888 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:24Z","lastTransitionTime":"2025-10-04T03:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.039352 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.039413 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.039430 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.039454 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.039471 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.141933 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.142355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.142549 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.142800 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.143044 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.246395 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.246461 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.246478 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.246503 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.246523 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.350062 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.350130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.350153 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.350182 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.350203 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.453257 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.453338 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.453357 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.453379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.453395 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.556701 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.556798 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.556827 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.556865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.556889 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.660578 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.660632 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.660649 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.660674 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.660691 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.673375 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.673409 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:25 crc kubenswrapper[4770]: E1004 03:04:25.673586 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:25 crc kubenswrapper[4770]: E1004 03:04:25.673722 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.764476 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.764541 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.764567 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.764597 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.764619 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.869525 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.869605 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.869631 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.869665 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.869690 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.973318 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.973396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.973418 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.973445 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:25 crc kubenswrapper[4770]: I1004 03:04:25.973463 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:25Z","lastTransitionTime":"2025-10-04T03:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.076872 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.076934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.076954 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.076980 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.076998 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:26Z","lastTransitionTime":"2025-10-04T03:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.180502 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.180581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.180598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.180626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.180648 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:26Z","lastTransitionTime":"2025-10-04T03:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.284097 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.284183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.284204 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.284232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.284254 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:26Z","lastTransitionTime":"2025-10-04T03:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.387451 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.387521 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.387533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.387554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.387569 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:26Z","lastTransitionTime":"2025-10-04T03:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.492310 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.492402 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.492418 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.492438 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.492463 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:26Z","lastTransitionTime":"2025-10-04T03:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.595941 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.596044 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.596064 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.596087 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.596104 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:26Z","lastTransitionTime":"2025-10-04T03:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.673326 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.673366 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:26 crc kubenswrapper[4770]: E1004 03:04:26.673503 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:26 crc kubenswrapper[4770]: E1004 03:04:26.673639 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.698525 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.698596 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.698615 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.698642 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.698662 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:26Z","lastTransitionTime":"2025-10-04T03:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.802098 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.802156 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.802167 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.802184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.802200 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:26Z","lastTransitionTime":"2025-10-04T03:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.905759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.905823 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.905834 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.905848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:26 crc kubenswrapper[4770]: I1004 03:04:26.905878 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:26Z","lastTransitionTime":"2025-10-04T03:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.008451 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.008516 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.008536 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.008567 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.008587 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.112589 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.112636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.112648 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.112667 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.112679 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.216701 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.216782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.216809 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.216840 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.216858 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.320457 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.320558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.320579 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.320604 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.320621 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.423838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.423934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.423953 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.423985 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.424027 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.526962 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.527052 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.527087 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.527115 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.527134 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.634850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.634916 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.634936 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.634967 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.634989 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.673078 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:27 crc kubenswrapper[4770]: E1004 03:04:27.673351 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.673414 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:27 crc kubenswrapper[4770]: E1004 03:04:27.673662 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.738224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.738393 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.738414 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.738436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.738455 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.841585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.841748 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.841769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.841797 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.841817 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.945142 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.945206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.945218 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.945243 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:27 crc kubenswrapper[4770]: I1004 03:04:27.945260 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:27Z","lastTransitionTime":"2025-10-04T03:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.049465 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.049520 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.049541 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.049565 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.049583 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.065779 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.065837 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.065850 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.065870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.065885 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: E1004 03:04:28.079659 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:28Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.083477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.083535 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.083548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.083565 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.083576 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: E1004 03:04:28.098054 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:28Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.102106 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.102143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.102158 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.102178 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.102191 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: E1004 03:04:28.114202 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:28Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.118653 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.118715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.118732 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.118759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.118779 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: E1004 03:04:28.140073 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:28Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.144188 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.144221 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.144229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.144244 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.144268 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: E1004 03:04:28.157413 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:28Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:28 crc kubenswrapper[4770]: E1004 03:04:28.157516 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.158686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.158729 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.158737 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.158754 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.158764 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.261214 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.261282 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.261300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.261324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.261345 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.364746 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.364893 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.364980 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.365083 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.365114 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.468978 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.469143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.469167 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.469191 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.469207 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.572392 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.572458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.572506 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.572531 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.572548 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.673617 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.673750 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:28 crc kubenswrapper[4770]: E1004 03:04:28.674512 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:28 crc kubenswrapper[4770]: E1004 03:04:28.675092 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.675234 4770 scope.go:117] "RemoveContainer" containerID="28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.676667 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.676719 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.676742 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.676770 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.676794 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.780641 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.780713 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.780726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.780746 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.780758 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.884852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.884887 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.884898 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.884933 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.884946 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.988924 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.988974 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.988988 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.989033 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:28 crc kubenswrapper[4770]: I1004 03:04:28.989051 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:28Z","lastTransitionTime":"2025-10-04T03:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.091876 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.091922 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.091940 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.091965 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.091982 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:29Z","lastTransitionTime":"2025-10-04T03:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.195037 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.195556 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.195569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.195588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.195601 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:29Z","lastTransitionTime":"2025-10-04T03:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.259071 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/2.log" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.262080 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.262680 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.294990 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.297795 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.297839 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.297852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.297871 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.297883 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:29Z","lastTransitionTime":"2025-10-04T03:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.326020 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:20Z\\\",\\\"message\\\":\\\"2025-10-04T03:03:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3\\\\n2025-10-04T03:03:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3 to /host/opt/cni/bin/\\\\n2025-10-04T03:03:34Z [verbose] multus-daemon started\\\\n2025-10-04T03:03:34Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:04:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:04:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.398146 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.400245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.400279 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.400290 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.400307 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.400319 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:29Z","lastTransitionTime":"2025-10-04T03:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.415130 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.430302 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.443470 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.453331 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.465990 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.478316 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.489650 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.505187 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.505235 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.505249 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.505266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.505280 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:29Z","lastTransitionTime":"2025-10-04T03:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.511494 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.525746 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.538285 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335abf46-3919-45f7-a60e-6a334f91853e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc7d55b1cdd86b702936e172a73c1e0f947f6743583a3cb55c02de151a59629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.558156 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.569604 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.585981 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.596251 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.607152 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.607585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.607617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.607627 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.607642 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.607652 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:29Z","lastTransitionTime":"2025-10-04T03:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.617352 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.672983 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.673089 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:29 crc kubenswrapper[4770]: E1004 03:04:29.673169 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:29 crc kubenswrapper[4770]: E1004 03:04:29.673260 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.709635 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.709699 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.709711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.709729 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.709760 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:29Z","lastTransitionTime":"2025-10-04T03:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.723516 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.747538 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:20Z\\\",\\\"message\\\":\\\"2025-10-04T03:03:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3\\\\n2025-10-04T03:03:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3 to /host/opt/cni/bin/\\\\n2025-10-04T03:03:34Z [verbose] multus-daemon started\\\\n2025-10-04T03:03:34Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:04:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:04:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.766020 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.783289 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.794529 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.806195 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.812809 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.812843 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.812852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.812865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.812873 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:29Z","lastTransitionTime":"2025-10-04T03:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.820625 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.834094 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.844047 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.855456 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.872941 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.882053 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.890309 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.897378 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335abf46-3919-45f7-a60e-6a334f91853e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc7d55b1cdd86b702936e172a73c1e0f947f6743583a3cb55c02de151a59629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.912544 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.915293 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.915317 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.915327 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.915340 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.915349 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:29Z","lastTransitionTime":"2025-10-04T03:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.923310 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.935158 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.947076 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:29 crc kubenswrapper[4770]: I1004 03:04:29.958579 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:29Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.017979 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.018127 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.018150 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.018169 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.018180 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.120557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.120611 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.120628 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.120651 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.120668 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.223878 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.223957 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.223982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.224046 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.224067 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.269106 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/3.log" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.270525 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/2.log" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.274929 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653" exitCode=1 Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.274966 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.274999 4770 scope.go:117] "RemoveContainer" containerID="28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.276194 4770 scope.go:117] "RemoveContainer" containerID="fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653" Oct 04 03:04:30 crc kubenswrapper[4770]: E1004 03:04:30.277324 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.303101 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.324481 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:20Z\\\",\\\"message\\\":\\\"2025-10-04T03:03:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3\\\\n2025-10-04T03:03:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3 to /host/opt/cni/bin/\\\\n2025-10-04T03:03:34Z [verbose] multus-daemon started\\\\n2025-10-04T03:03:34Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:04:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:04:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.326248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.326271 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.326279 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.326292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.326301 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.342926 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.361919 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.378146 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.394807 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.406954 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.419613 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.430225 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.430583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.431265 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.431563 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.431719 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.432188 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.446481 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.475989 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d3a467b5204513c5ecf8f4772683e56fa71b7af79b55744a40c95cf826491f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:02Z\\\",\\\"message\\\":\\\"}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:04:02.659334 6437 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1004 03:04:02.660145 6437 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-controller template LB for network=default: []services.LB{}\\\\nI1004 03:04:02.660240 6437 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 1.274544ms)\\\\nI1004 03:04:02.660310 6437 services_controller.go:454] Service openshift-machine-config-operator/machine-config-controller for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 03:04:02.659308 6437 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:04:02.660362 6437 factory.go:656] Stopping watch factory\\\\nI1004 03:04:02.660373 6437 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:04:02.660402 6437 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:04:02.660414 6437 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:04:02.660476 6437 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:29Z\\\",\\\"message\\\":\\\"9.784330 6774 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1004 03:04:29.784333 6774 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 03:04:29.784177 6774 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-26dzh\\\\nI1004 03:04:29.784345 6774 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 03:04:29.784287 6774 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1004 03:04:29.784356 6774 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1004 03:04:29.784363 6774 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1004 03:04:29.784396 6774 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed t\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.493070 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.540349 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.540415 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.540433 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.540460 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.540480 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.549300 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335abf46-3919-45f7-a60e-6a334f91853e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc7d55b1cdd86b702936e172a73c1e0f947f6743583a3cb55c02de151a59629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.570208 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.583685 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.595402 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.607553 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.619739 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.631487 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:30Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.642538 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.642602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.642615 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.642634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.642647 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.673034 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:30 crc kubenswrapper[4770]: E1004 03:04:30.673149 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.673330 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:30 crc kubenswrapper[4770]: E1004 03:04:30.673648 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.745938 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.746319 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.746345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.746424 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.746446 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.849462 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.849492 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.849502 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.849517 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.849526 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.952432 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.952484 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.952502 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.952526 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:30 crc kubenswrapper[4770]: I1004 03:04:30.952544 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:30Z","lastTransitionTime":"2025-10-04T03:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.055281 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.055335 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.055356 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.055381 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.055396 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.157222 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.157264 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.157275 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.157291 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.157303 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.259918 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.259946 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.259954 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.259965 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.259974 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.279478 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/3.log" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.282907 4770 scope.go:117] "RemoveContainer" containerID="fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653" Oct 04 03:04:31 crc kubenswrapper[4770]: E1004 03:04:31.283045 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.299232 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.317735 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:20Z\\\",\\\"message\\\":\\\"2025-10-04T03:03:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3\\\\n2025-10-04T03:03:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3 to /host/opt/cni/bin/\\\\n2025-10-04T03:03:34Z [verbose] multus-daemon started\\\\n2025-10-04T03:03:34Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:04:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:04:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.341070 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.357775 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.362869 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.362909 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.362918 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.362934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.362945 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.373466 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.396042 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.414451 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.433520 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.449175 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.466177 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.466231 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.466246 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.466266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.466282 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.467253 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.504844 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:29Z\\\",\\\"message\\\":\\\"9.784330 6774 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1004 03:04:29.784333 6774 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 03:04:29.784177 6774 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-26dzh\\\\nI1004 03:04:29.784345 6774 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 03:04:29.784287 6774 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1004 03:04:29.784356 6774 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1004 03:04:29.784363 6774 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1004 03:04:29.784396 6774 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed t\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.521736 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.540621 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.563320 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.569820 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.569924 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.569944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.569967 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.569988 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.580566 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.596276 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335abf46-3919-45f7-a60e-6a334f91853e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc7d55b1cdd86b702936e172a73c1e0f947f6743583a3cb55c02de151a59629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.622662 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.643748 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.663735 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.672672 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:31 crc kubenswrapper[4770]: E1004 03:04:31.672815 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.672918 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:31 crc kubenswrapper[4770]: E1004 03:04:31.673055 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.673292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.673339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.673365 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.673399 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.673426 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.777429 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.777479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.777491 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.777511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.777526 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.880363 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.880865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.880960 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.881066 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.881152 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.984045 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.984073 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.984081 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.984094 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:31 crc kubenswrapper[4770]: I1004 03:04:31.984103 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:31Z","lastTransitionTime":"2025-10-04T03:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.087891 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.087955 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.087975 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.088001 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.088055 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:32Z","lastTransitionTime":"2025-10-04T03:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.192123 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.192181 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.192192 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.192213 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.192224 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:32Z","lastTransitionTime":"2025-10-04T03:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.294513 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.294543 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.294551 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.294564 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.294573 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:32Z","lastTransitionTime":"2025-10-04T03:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.397962 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.398818 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.399119 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.399300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.399479 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:32Z","lastTransitionTime":"2025-10-04T03:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.503185 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.503259 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.503279 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.503304 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.503323 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:32Z","lastTransitionTime":"2025-10-04T03:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.607371 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.607416 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.607434 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.607457 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.607475 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:32Z","lastTransitionTime":"2025-10-04T03:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.673280 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.673398 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:32 crc kubenswrapper[4770]: E1004 03:04:32.673486 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:32 crc kubenswrapper[4770]: E1004 03:04:32.673829 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.710496 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.710936 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.711053 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.711141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.711205 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:32Z","lastTransitionTime":"2025-10-04T03:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.814460 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.814513 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.814530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.814555 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.814573 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:32Z","lastTransitionTime":"2025-10-04T03:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.918417 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.918482 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.918502 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.918526 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:32 crc kubenswrapper[4770]: I1004 03:04:32.918546 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:32Z","lastTransitionTime":"2025-10-04T03:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.021721 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.021783 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.021810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.021845 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.021871 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.125358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.125406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.125423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.125449 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.125463 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.229185 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.229251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.229270 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.229294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.229312 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.332232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.332290 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.332309 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.332336 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.332355 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.436527 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.436591 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.436604 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.436622 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.436633 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.539633 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.539691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.539705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.539730 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.539745 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.643430 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.643489 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.643501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.643520 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.643533 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.672851 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.672925 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:33 crc kubenswrapper[4770]: E1004 03:04:33.673090 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:33 crc kubenswrapper[4770]: E1004 03:04:33.673335 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.746772 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.746848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.746863 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.746888 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.746907 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.849893 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.849965 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.849983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.850055 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.850075 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.954258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.954334 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.954354 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.954420 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:33 crc kubenswrapper[4770]: I1004 03:04:33.954438 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:33Z","lastTransitionTime":"2025-10-04T03:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.057711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.057773 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.057787 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.057817 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.057833 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.161806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.161882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.161902 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.161931 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.161953 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.267065 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.267448 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.267542 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.267610 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.267736 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.372134 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.372403 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.372429 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.372466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.372499 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.375413 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.375721 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.375677693 +0000 UTC m=+149.667687445 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.475749 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.475801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.475815 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.475834 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.475845 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.476197 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.476241 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.476267 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.476295 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476378 4770 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476418 4770 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476438 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.476418024 +0000 UTC m=+149.768427736 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476469 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476493 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.476469236 +0000 UTC m=+149.768479188 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476495 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476517 4770 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476542 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476584 4770 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476604 4770 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476557 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.476546148 +0000 UTC m=+149.768556100 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.476764 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.476692221 +0000 UTC m=+149.768701943 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.580075 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.580184 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.580204 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.580267 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.580287 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.673026 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.673037 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.673255 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:34 crc kubenswrapper[4770]: E1004 03:04:34.673385 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.683438 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.683501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.683525 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.683557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.683579 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.787411 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.787471 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.787487 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.787510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.787527 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.890620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.890656 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.890665 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.890680 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.890691 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.993772 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.993848 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.993870 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.993901 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:34 crc kubenswrapper[4770]: I1004 03:04:34.993924 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:34Z","lastTransitionTime":"2025-10-04T03:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.097617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.097711 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.097737 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.097769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.097788 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:35Z","lastTransitionTime":"2025-10-04T03:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.202868 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.202933 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.202951 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.202979 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.203002 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:35Z","lastTransitionTime":"2025-10-04T03:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.306413 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.306485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.306508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.306538 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.306557 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:35Z","lastTransitionTime":"2025-10-04T03:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.409858 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.409932 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.409951 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.409982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.410003 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:35Z","lastTransitionTime":"2025-10-04T03:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.515224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.515300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.515324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.515367 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.515395 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:35Z","lastTransitionTime":"2025-10-04T03:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.619714 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.619793 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.619818 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.619852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.619880 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:35Z","lastTransitionTime":"2025-10-04T03:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.674340 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.674623 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:35 crc kubenswrapper[4770]: E1004 03:04:35.674969 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:35 crc kubenswrapper[4770]: E1004 03:04:35.675370 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.728271 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.728608 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.728983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.729054 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.729075 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:35Z","lastTransitionTime":"2025-10-04T03:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.833188 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.833267 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.833283 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.833303 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.833345 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:35Z","lastTransitionTime":"2025-10-04T03:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.937050 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.937100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.937116 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.937140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:35 crc kubenswrapper[4770]: I1004 03:04:35.937158 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:35Z","lastTransitionTime":"2025-10-04T03:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.040581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.041085 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.041299 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.041448 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.041592 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.146210 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.146268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.146287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.146313 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.146333 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.249353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.249400 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.249413 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.249431 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.249444 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.352455 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.352496 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.352507 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.352523 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.352535 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.455397 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.455892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.455906 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.455922 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.455935 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.559490 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.559565 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.559589 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.559617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.559638 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.662891 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.662959 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.662976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.662998 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.663045 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.673544 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.673675 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:36 crc kubenswrapper[4770]: E1004 03:04:36.673744 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:36 crc kubenswrapper[4770]: E1004 03:04:36.673884 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.766343 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.766399 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.766419 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.766446 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.766464 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.869419 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.869469 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.869479 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.869499 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.869511 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.973201 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.973271 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.973292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.973321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:36 crc kubenswrapper[4770]: I1004 03:04:36.973340 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:36Z","lastTransitionTime":"2025-10-04T03:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.078324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.078388 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.078414 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.078451 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.078478 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:37Z","lastTransitionTime":"2025-10-04T03:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.182414 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.182523 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.182550 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.182590 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.182617 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:37Z","lastTransitionTime":"2025-10-04T03:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.289376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.289443 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.289462 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.289497 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.289514 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:37Z","lastTransitionTime":"2025-10-04T03:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.392865 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.392914 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.392927 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.392944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.392956 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:37Z","lastTransitionTime":"2025-10-04T03:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.498046 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.498101 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.498120 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.498146 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.498165 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:37Z","lastTransitionTime":"2025-10-04T03:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.603745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.603819 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.603838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.603866 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.603888 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:37Z","lastTransitionTime":"2025-10-04T03:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.673141 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.673220 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:37 crc kubenswrapper[4770]: E1004 03:04:37.673476 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:37 crc kubenswrapper[4770]: E1004 03:04:37.673669 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.709483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.709568 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.709595 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.709634 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.709659 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:37Z","lastTransitionTime":"2025-10-04T03:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.812421 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.812481 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.812498 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.812516 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.812528 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:37Z","lastTransitionTime":"2025-10-04T03:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.916722 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.916786 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.916806 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.916830 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:37 crc kubenswrapper[4770]: I1004 03:04:37.916844 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:37Z","lastTransitionTime":"2025-10-04T03:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.020679 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.020739 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.020760 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.020787 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.020809 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.124192 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.124266 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.124287 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.124317 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.124336 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.227904 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.227964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.227982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.228035 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.228054 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.332038 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.332090 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.332100 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.332121 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.332134 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.380555 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.380626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.380645 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.380673 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.380695 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: E1004 03:04:38.402511 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.408777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.408844 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.408864 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.408892 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.408911 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: E1004 03:04:38.433073 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.439486 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.439536 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.439554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.439576 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.439591 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: E1004 03:04:38.455410 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.462180 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.462249 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.462264 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.462292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.462311 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: E1004 03:04:38.486080 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.492341 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.492418 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.492440 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.492470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.492494 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: E1004 03:04:38.514207 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:38 crc kubenswrapper[4770]: E1004 03:04:38.514481 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.517003 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.517110 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.517132 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.517163 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.517187 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.621602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.621669 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.621682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.621705 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.621721 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.672929 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.672929 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:38 crc kubenswrapper[4770]: E1004 03:04:38.673165 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:38 crc kubenswrapper[4770]: E1004 03:04:38.673267 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.726043 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.726097 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.726115 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.726139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.726157 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.829733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.829801 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.829814 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.829838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.829856 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.933516 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.933593 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.933610 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.933637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:38 crc kubenswrapper[4770]: I1004 03:04:38.933661 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:38Z","lastTransitionTime":"2025-10-04T03:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.037785 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.037879 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.037899 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.037930 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.037953 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.140829 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.140873 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.140882 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.140900 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.140913 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.244351 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.244423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.244442 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.244469 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.244491 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.347288 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.347349 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.347365 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.347391 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.347412 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.450170 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.450263 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.450292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.450325 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.450349 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.553150 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.553206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.553216 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.553245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.553260 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.656616 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.656714 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.656741 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.656775 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.656801 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.673190 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.673311 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:39 crc kubenswrapper[4770]: E1004 03:04:39.673458 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:39 crc kubenswrapper[4770]: E1004 03:04:39.673673 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.694574 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krp9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kv2f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krp9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.716316 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"335abf46-3919-45f7-a60e-6a334f91853e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc7d55b1cdd86b702936e172a73c1e0f947f6743583a3cb55c02de151a59629a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://907080b4c19134978f6f7134f353fab0fcbb03c21c829451694d16c1a10c008e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.744989 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bff5b753-afcb-475f-bfa7-c30b62b347a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd96ee6f9f2a833d2b499f7cb14677ae6cfd89f7a6c1bf58bf56c29e58592ff3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8366d4f633e9d38cbbe5e739fb0380884896d7a76161eba2b609d139fc33df9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dccd51de3c19b74439072ce0ffe0b6514c96e3c6ee20e4b0c62c1b9b14ecc89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69f8c45e78d18e2ac798cdf5bf8fb4ee9f2c757f4fb226cc8cdb2160e4f92860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://820a70b07f4b7f70d2df6e0c8d9b8c7eb9857eac2584ea55f7a1fb1a71b53280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51519fbe8cbf016eb1aabdf7b0ba54f04f69c6d6c035ba452258c5903a4d66fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec3719caad4a2e483b667f398a4b10d76e1e0f4ce6d44501b5d731664bde2c1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b8a0eee34cb58ec3479da78fba01bae7dbf0e33433681f9c35456cc0b1e56ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.760657 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.760712 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.760730 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.760756 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.760774 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.770304 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0323964d-7df2-4fc4-a8a2-e365d071a76b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f15c443304990d7b84918d6fe2c1992e71166ca6caf3ee9530a8108739abc877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f20ef83e3dfe2d0a188016e76ef9c00a0712396c82ae7c54000aef05217673\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b538e3eb24349135a0fae4cb14e6b0d8536cb90fb9f9c63eb4afbdb50c913b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b44a4dc86ddd93d1cee027bed1ccc61e9d25898e5409fcd20c99db9c0716af51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d2f6b913f27bcf2623d6cfaed3ff94c0596a4f1e000a3f4b146e7562f41545\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:03:30.595385 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:03:30.595522 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:03:30.596443 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2548621211/tls.crt::/tmp/serving-cert-2548621211/tls.key\\\\\\\"\\\\nI1004 03:03:31.053960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:03:31.056969 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:03:31.056988 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:03:31.057024 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:03:31.057031 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:03:31.069462 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:03:31.069515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:03:31.069538 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:03:31.069543 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:03:31.069547 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:03:31.069550 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:03:31.071401 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:03:31.071763 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://04244d162eaf46bf2dc1b20bcfae263aa019cc11dbbbf59484b930d90e543ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ff2e3723e47d0b7757b20f5ad5c796792235a69195f88428db46eef9e5bbfa8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.791087 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb37f3f2-7bdd-4d92-9595-c9d04579adc2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb8314be5ac06c5324e4a5cb23bc76ac6dcedf077691f643a9805a1085a6c07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c2cce1f55a8406f940621f5c7c758206fad8aefc35769b81a31d4d090acf255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d3fa822b358168eae1a0553614ebbd44d6aecdecf4637614ac2d1aa51fa7346\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab49494ae88337538a2d92edf554a9d64c3be9796263893ade6266b1e8131b64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.810708 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fa8c67-3d3f-4cfb-a023-72ed150ed6c0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357a260211478335b0cd32d605e7a4306494d9a7b11b81e60bc8e54bb9c9fc55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9e80433fa25c092c012cef905ab141e7308d59b22926027f5d8724bbfc58fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71458de19cca2d1af873bfe8a79e25ef27bf2911119d132be08dba73cc2479ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://273ab21e8b43d6aa79c51763f80574a33d457815a66132a6be59e0ca1cb9aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.829083 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0756b7ffce9d895f280dcf7e55bd3492db66b759b19ad217eedf3dd43596fdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e277a680f1b754f5b8164261f4264a3bc9c006481f514caec55fc96c26db1eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.848916 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.865196 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-66tn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88d62119-65f3-4846-8b93-53d096fa9df7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:20Z\\\",\\\"message\\\":\\\"2025-10-04T03:03:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3\\\\n2025-10-04T03:03:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27cba606-537b-4d34-8738-38f4480b40c3 to /host/opt/cni/bin/\\\\n2025-10-04T03:03:34Z [verbose] multus-daemon started\\\\n2025-10-04T03:03:34Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:04:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:04:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hm5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-66tn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.865378 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.865443 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.865471 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.865509 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.865535 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.888928 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26dzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc1d26ac-d5d4-4841-a45b-809a9ba074cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcef1a7d0e0080b64d5b8793c5910df38439ac73843d2447da81a7d15fcde393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cbeacd1eefd3caef4cb871b3a4f49c037a92a66920bc2e136d0a1d85dea442f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f50c6bb906c3966e001080ac80046d41b6761dc300ae6e08223942ed66a5b12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://619430ced04b8ec0e4ad04c87d49a3154d294d5f1cbd0050e5ba69312d744029\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://098dbc45e86773da267c19503cd9c2d0b5d34e7cadf32353504a0ee0553d10ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64233c96c6725ae0a0e76bf96e9d14bfa3d5ae162fa88cb7eb2a81c4bdfd8ef2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9546153664d180a3d48f2e563aaf26d257db627032d0ec83848bb7b2fbbd595a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pmrx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26dzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.906314 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ad38ecf62f06b5d7483ffd0d850cbf7ebc3f3ecb3d7fe21a8dce63d3d4119f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.926171 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.945687 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74c7f397b9afce4a7985ea36c7903b494eaf57236d58716d2dd86025a771eaff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.961510 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xsrdd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a928bfe-7bcc-4d20-9eda-f61f1d0b7733\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a952459dda85adeef95fa52e67874f73e93e3a79d1f85dd1ef6d656b4acdb053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlr9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xsrdd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.968516 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.968567 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.968594 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.968621 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.968639 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:39Z","lastTransitionTime":"2025-10-04T03:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.979453 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6bd3be93-1791-4cd1-b3ae-b4032548e93a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9e1c1c6d75a22440967afb07f8e82b4dce6cd1d3f609da97402a1b9e851e7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gxsxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-h6msx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:39 crc kubenswrapper[4770]: I1004 03:04:39.996210 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7wlm6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d8687a3-eceb-430c-aed8-ee3bb1493f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c451eeb09bfc28b24b441617839eaf78f72d51e0271b8440ae25f1643611fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zjkwq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7wlm6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:39Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.016969 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.049258 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:29Z\\\",\\\"message\\\":\\\"9.784330 6774 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1004 03:04:29.784333 6774 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 03:04:29.784177 6774 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-26dzh\\\\nI1004 03:04:29.784345 6774 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 03:04:29.784287 6774 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1004 03:04:29.784356 6774 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1004 03:04:29.784363 6774 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1004 03:04:29.784396 6774 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed t\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.066976 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:40Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.071488 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.071530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.071547 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.071571 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.071589 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:40Z","lastTransitionTime":"2025-10-04T03:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.174336 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.174408 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.174436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.174472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.174498 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:40Z","lastTransitionTime":"2025-10-04T03:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.278501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.278598 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.278682 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.278712 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.278731 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:40Z","lastTransitionTime":"2025-10-04T03:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.383466 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.383530 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.383544 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.383569 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.383588 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:40Z","lastTransitionTime":"2025-10-04T03:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.486914 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.487107 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.487129 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.487154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.487171 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:40Z","lastTransitionTime":"2025-10-04T03:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.591472 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.591570 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.591582 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.591601 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.591618 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:40Z","lastTransitionTime":"2025-10-04T03:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.673160 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.673270 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:40 crc kubenswrapper[4770]: E1004 03:04:40.673406 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:40 crc kubenswrapper[4770]: E1004 03:04:40.673642 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.695138 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.695222 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.695240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.695270 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.695300 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:40Z","lastTransitionTime":"2025-10-04T03:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.798605 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.798679 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.798709 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.798745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.798773 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:40Z","lastTransitionTime":"2025-10-04T03:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.902557 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.902651 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.902670 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.902709 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:40 crc kubenswrapper[4770]: I1004 03:04:40.902731 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:40Z","lastTransitionTime":"2025-10-04T03:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.006204 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.006297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.006318 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.006349 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.006368 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.110305 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.110384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.110408 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.110439 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.110467 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.221306 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.221420 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.221474 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.221702 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.221790 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.326203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.326251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.326262 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.326277 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.326288 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.430129 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.430188 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.430206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.430227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.430244 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.533927 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.534065 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.534096 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.534142 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.534172 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.638130 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.638205 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.638224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.638255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.638274 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.673523 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.673727 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:41 crc kubenswrapper[4770]: E1004 03:04:41.673891 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:41 crc kubenswrapper[4770]: E1004 03:04:41.674187 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.742247 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.742309 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.742328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.742353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.742372 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.846065 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.846138 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.846157 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.846182 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.846201 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.949959 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.950090 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.950117 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.950151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:41 crc kubenswrapper[4770]: I1004 03:04:41.950174 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:41Z","lastTransitionTime":"2025-10-04T03:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.054520 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.054586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.054605 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.054631 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.054649 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.158867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.158930 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.158947 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.158974 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.158992 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.262857 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.262928 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.262948 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.262979 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.263035 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.366499 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.366620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.366648 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.366684 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.366893 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.471105 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.471151 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.471161 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.471180 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.471195 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.575255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.575331 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.575349 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.575379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.575396 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.673261 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.673325 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:42 crc kubenswrapper[4770]: E1004 03:04:42.673461 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:42 crc kubenswrapper[4770]: E1004 03:04:42.673609 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.674704 4770 scope.go:117] "RemoveContainer" containerID="fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653" Oct 04 03:04:42 crc kubenswrapper[4770]: E1004 03:04:42.675001 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.679458 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.679536 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.679558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.679588 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.679614 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.782617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.782668 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.782678 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.782697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.782708 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.886889 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.886950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.886963 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.886983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.886994 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.990700 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.990749 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.990764 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.990782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:42 crc kubenswrapper[4770]: I1004 03:04:42.990986 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:42Z","lastTransitionTime":"2025-10-04T03:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.094690 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.094726 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.094735 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.094748 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.094757 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:43Z","lastTransitionTime":"2025-10-04T03:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.198423 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.198510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.198535 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.198566 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.198590 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:43Z","lastTransitionTime":"2025-10-04T03:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.301984 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.302095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.302120 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.302147 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.302170 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:43Z","lastTransitionTime":"2025-10-04T03:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.405625 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.405667 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.405679 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.405697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.405711 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:43Z","lastTransitionTime":"2025-10-04T03:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.509039 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.509069 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.509078 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.509091 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.509100 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:43Z","lastTransitionTime":"2025-10-04T03:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.613396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.613471 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.613484 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.613505 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.613522 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:43Z","lastTransitionTime":"2025-10-04T03:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.673318 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.673380 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:43 crc kubenswrapper[4770]: E1004 03:04:43.673472 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:43 crc kubenswrapper[4770]: E1004 03:04:43.673738 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.717389 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.717454 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.717465 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.717487 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.717502 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:43Z","lastTransitionTime":"2025-10-04T03:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.821558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.821613 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.821632 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.821662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.821695 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:43Z","lastTransitionTime":"2025-10-04T03:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.925157 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.925209 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.925227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.925253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:43 crc kubenswrapper[4770]: I1004 03:04:43.925273 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:43Z","lastTransitionTime":"2025-10-04T03:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.029410 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.029494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.029508 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.029532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.029544 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.133188 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.133251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.133268 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.133292 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.133357 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.236597 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.236677 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.236733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.236770 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.236791 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.339702 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.339758 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.339776 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.339802 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.339838 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.442939 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.442993 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.443045 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.443074 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.443092 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.546194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.546267 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.546291 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.546318 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.546344 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.650032 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.650096 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.650113 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.650144 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.650183 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.672744 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:44 crc kubenswrapper[4770]: E1004 03:04:44.672995 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.673296 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:44 crc kubenswrapper[4770]: E1004 03:04:44.673634 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.753340 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.753390 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.753406 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.753426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.753440 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.856549 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.856626 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.856637 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.856658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.856669 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.960293 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.960362 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.960387 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.960420 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:44 crc kubenswrapper[4770]: I1004 03:04:44.960443 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:44Z","lastTransitionTime":"2025-10-04T03:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.063446 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.063910 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.064048 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.064220 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.064356 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.168049 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.168137 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.168157 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.168183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.168201 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.271264 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.271308 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.271321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.271339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.271352 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.374464 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.374535 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.374555 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.374586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.374604 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.478273 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.478362 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.478388 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.478419 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.478437 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.581374 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.581473 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.581498 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.581527 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.581548 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.672727 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.672751 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:45 crc kubenswrapper[4770]: E1004 03:04:45.672988 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:45 crc kubenswrapper[4770]: E1004 03:04:45.673106 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.684824 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.684883 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.684909 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.684940 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.684963 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.788599 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.788698 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.788715 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.788745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.788764 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.895532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.895663 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.895686 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.895716 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.895747 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.999470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.999542 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.999559 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.999587 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:45 crc kubenswrapper[4770]: I1004 03:04:45.999606 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:45Z","lastTransitionTime":"2025-10-04T03:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.102494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.102561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.102580 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.102606 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.102625 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:46Z","lastTransitionTime":"2025-10-04T03:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.205929 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.206305 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.206509 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.206906 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.207256 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:46Z","lastTransitionTime":"2025-10-04T03:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.310230 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.310572 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.310908 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.311093 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.311265 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:46Z","lastTransitionTime":"2025-10-04T03:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.414616 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.414688 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.414706 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.414733 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.414751 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:46Z","lastTransitionTime":"2025-10-04T03:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.518345 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.518703 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.518791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.518886 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.518967 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:46Z","lastTransitionTime":"2025-10-04T03:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.621640 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.621980 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.622086 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.622180 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.622284 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:46Z","lastTransitionTime":"2025-10-04T03:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.673271 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.673314 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:46 crc kubenswrapper[4770]: E1004 03:04:46.673779 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:46 crc kubenswrapper[4770]: E1004 03:04:46.674203 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.725983 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.726324 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.726507 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.726689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.726818 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:46Z","lastTransitionTime":"2025-10-04T03:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.830981 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.831083 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.831106 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.831164 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.831183 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:46Z","lastTransitionTime":"2025-10-04T03:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.935194 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.935240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.935256 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.935280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:46 crc kubenswrapper[4770]: I1004 03:04:46.935297 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:46Z","lastTransitionTime":"2025-10-04T03:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.038944 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.039224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.039251 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.039283 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.039308 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.142576 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.142654 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.142680 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.142758 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.142819 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.245753 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.246290 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.246460 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.246594 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.246723 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.348836 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.349106 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.349201 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.349441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.349542 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.452782 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.453248 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.453390 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.453531 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.453662 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.557341 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.557416 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.557436 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.557470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.557491 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.661687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.662213 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.662433 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.662687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.662893 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.673236 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.673368 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:47 crc kubenswrapper[4770]: E1004 03:04:47.673663 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:47 crc kubenswrapper[4770]: E1004 03:04:47.673929 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.766365 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.766427 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.766446 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.766473 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.766492 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.869422 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.869489 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.869511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.869536 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.869553 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.973649 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.973704 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.973717 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.973739 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:47 crc kubenswrapper[4770]: I1004 03:04:47.973755 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:47Z","lastTransitionTime":"2025-10-04T03:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.076687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.076766 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.076786 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.076810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.076829 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.180252 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.180331 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.180353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.180376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.180392 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.283981 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.284095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.284114 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.284139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.284159 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.388519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.388602 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.388619 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.388654 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.388675 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.501234 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.501297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.501315 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.501339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.501356 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.605523 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.605585 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.605605 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.605631 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.605652 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.673424 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.673410 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:48 crc kubenswrapper[4770]: E1004 03:04:48.674218 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:48 crc kubenswrapper[4770]: E1004 03:04:48.674257 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.709376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.709788 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.709955 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.710169 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.710345 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.814140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.814189 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.814199 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.814219 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.814231 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.863989 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.864110 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.864124 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.864141 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.864151 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: E1004 03:04:48.883448 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.889238 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.889294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.889310 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.889330 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.889345 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: E1004 03:04:48.907701 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.912562 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.912627 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.912636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.912652 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.912662 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: E1004 03:04:48.927438 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.931584 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.931620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.931632 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.931646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.931657 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: E1004 03:04:48.950197 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.955444 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.955481 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.955493 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.955510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.955521 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:48 crc kubenswrapper[4770]: E1004 03:04:48.972223 4770 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9ca6cfa6-9f62-40aa-a5f8-5a3e3ff906b8\\\",\\\"systemUUID\\\":\\\"3e1fbf36-5852-424f-a63e-18ade8ba99b6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:48 crc kubenswrapper[4770]: E1004 03:04:48.972392 4770 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.974269 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.974326 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.974339 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.974358 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:48 crc kubenswrapper[4770]: I1004 03:04:48.974371 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:48Z","lastTransitionTime":"2025-10-04T03:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.077162 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.077238 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.077278 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.077308 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.077330 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:49Z","lastTransitionTime":"2025-10-04T03:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.180126 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.180183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.180193 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.180206 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.180215 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:49Z","lastTransitionTime":"2025-10-04T03:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.284028 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.284087 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.284099 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.284115 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.284125 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:49Z","lastTransitionTime":"2025-10-04T03:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.387575 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.387650 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.387661 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.387676 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.387686 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:49Z","lastTransitionTime":"2025-10-04T03:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.490759 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.490852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.490874 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.490905 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.490924 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:49Z","lastTransitionTime":"2025-10-04T03:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.595139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.595222 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.595242 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.595274 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.595297 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:49Z","lastTransitionTime":"2025-10-04T03:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.673553 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.673899 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:49 crc kubenswrapper[4770]: E1004 03:04:49.674316 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:49 crc kubenswrapper[4770]: E1004 03:04:49.674539 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.700823 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.701433 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.701484 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.701493 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.701513 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.701524 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:49Z","lastTransitionTime":"2025-10-04T03:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.739533 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:04:29Z\\\",\\\"message\\\":\\\"9.784330 6774 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1004 03:04:29.784333 6774 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 03:04:29.784177 6774 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-26dzh\\\\nI1004 03:04:29.784345 6774 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 03:04:29.784287 6774 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1004 03:04:29.784356 6774 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1004 03:04:29.784363 6774 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1004 03:04:29.784396 6774 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed t\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:04:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:03:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:03:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg9hp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rtksw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.757828 4770 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3c536f-6774-43e6-8800-e7e060cd96f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:03:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b506a09a154c71b66e93c1b74884f4ca0f010f9af3e7bded7af38f1001402d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f9a7274d5e44835156e23977ff8061c217d65df796710677ae1160c13682bda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whw6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:03:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-z9xqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:04:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.816233 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.816297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.816308 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.816330 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.816361 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:49Z","lastTransitionTime":"2025-10-04T03:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.818478 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.818459417 podStartE2EDuration="25.818459417s" podCreationTimestamp="2025-10-04 03:04:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:49.784825635 +0000 UTC m=+101.076835347" watchObservedRunningTime="2025-10-04 03:04:49.818459417 +0000 UTC m=+101.110469129" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.838835 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=79.838814455 podStartE2EDuration="1m19.838814455s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:49.819374991 +0000 UTC m=+101.111384723" watchObservedRunningTime="2025-10-04 03:04:49.838814455 +0000 UTC m=+101.130824167" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.856888 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.856857773 podStartE2EDuration="1m18.856857773s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:49.839732029 +0000 UTC m=+101.131741761" watchObservedRunningTime="2025-10-04 03:04:49.856857773 +0000 UTC m=+101.148867485" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.859591 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:49 crc kubenswrapper[4770]: E1004 03:04:49.859929 4770 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:04:49 crc kubenswrapper[4770]: E1004 03:04:49.860035 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs podName:5e7fe8aa-4b62-47da-8aa7-0d31dcdba457 nodeName:}" failed. No retries permitted until 2025-10-04 03:05:53.859996884 +0000 UTC m=+165.152006596 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs") pod "network-metrics-daemon-krp9c" (UID: "5e7fe8aa-4b62-47da-8aa7-0d31dcdba457") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.887402 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.887368444 podStartE2EDuration="1m15.887368444s" podCreationTimestamp="2025-10-04 03:03:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:49.856967285 +0000 UTC m=+101.148977057" watchObservedRunningTime="2025-10-04 03:04:49.887368444 +0000 UTC m=+101.179378176" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.887644 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.887636081 podStartE2EDuration="52.887636081s" podCreationTimestamp="2025-10-04 03:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:49.887189978 +0000 UTC m=+101.179199690" watchObservedRunningTime="2025-10-04 03:04:49.887636081 +0000 UTC m=+101.179645813" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.922930 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.922978 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.922988 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.923027 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.923040 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:49Z","lastTransitionTime":"2025-10-04T03:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.962056 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-66tn4" podStartSLOduration=79.961994178 podStartE2EDuration="1m19.961994178s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:49.959229246 +0000 UTC m=+101.251238978" watchObservedRunningTime="2025-10-04 03:04:49.961994178 +0000 UTC m=+101.254003910" Oct 04 03:04:49 crc kubenswrapper[4770]: I1004 03:04:49.992967 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-26dzh" podStartSLOduration=79.992935309 podStartE2EDuration="1m19.992935309s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:49.977303694 +0000 UTC m=+101.269313416" watchObservedRunningTime="2025-10-04 03:04:49.992935309 +0000 UTC m=+101.284945041" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.026617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.027043 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.027258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.027416 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.027500 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.071999 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podStartSLOduration=80.071975118 podStartE2EDuration="1m20.071975118s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:50.071640849 +0000 UTC m=+101.363650571" watchObservedRunningTime="2025-10-04 03:04:50.071975118 +0000 UTC m=+101.363984840" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.072674 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-xsrdd" podStartSLOduration=80.072669616 podStartE2EDuration="1m20.072669616s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:50.050273346 +0000 UTC m=+101.342283058" watchObservedRunningTime="2025-10-04 03:04:50.072669616 +0000 UTC m=+101.364679328" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.087622 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7wlm6" podStartSLOduration=79.087562983 podStartE2EDuration="1m19.087562983s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:50.086980737 +0000 UTC m=+101.378990459" watchObservedRunningTime="2025-10-04 03:04:50.087562983 +0000 UTC m=+101.379572695" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.131685 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.131737 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.131755 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.131777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.131793 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.234224 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.234282 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.234297 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.234321 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.234337 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.338667 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.338779 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.338791 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.338814 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.338828 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.441845 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.442031 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.442055 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.442109 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.442238 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.546667 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.546745 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.546768 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.546846 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.546866 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.650197 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.650300 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.650319 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.650355 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.650379 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.673051 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.673148 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:50 crc kubenswrapper[4770]: E1004 03:04:50.673273 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:50 crc kubenswrapper[4770]: E1004 03:04:50.673576 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.753803 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.753875 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.753891 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.753919 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.753936 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.857903 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.857984 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.858032 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.858066 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.858088 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.960813 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.960858 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.960867 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.960881 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:50 crc kubenswrapper[4770]: I1004 03:04:50.960891 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:50Z","lastTransitionTime":"2025-10-04T03:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.063914 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.063958 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.063970 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.063989 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.064006 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.168186 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.168245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.168258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.168282 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.168299 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.271691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.271788 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.271808 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.271838 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.271857 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.374852 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.374953 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.374979 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.375043 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.375070 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.478101 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.478162 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.478174 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.478203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.478216 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.581673 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.581751 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.581777 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.581807 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.581830 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.673819 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.673843 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:51 crc kubenswrapper[4770]: E1004 03:04:51.674257 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:51 crc kubenswrapper[4770]: E1004 03:04:51.674456 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.684564 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.684629 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.684650 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.684679 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.684700 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.788891 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.788958 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.788976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.789034 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.789053 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.892450 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.892533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.892591 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.892620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.892640 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.995500 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.996168 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.996211 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.996234 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:51 crc kubenswrapper[4770]: I1004 03:04:51.996248 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:51Z","lastTransitionTime":"2025-10-04T03:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.099813 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.099881 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.099904 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.099937 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.099964 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:52Z","lastTransitionTime":"2025-10-04T03:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.203989 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.204095 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.204299 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.204326 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.204349 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:52Z","lastTransitionTime":"2025-10-04T03:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.308426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.308510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.308534 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.308575 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.308600 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:52Z","lastTransitionTime":"2025-10-04T03:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.412401 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.412470 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.412485 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.412510 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.412525 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:52Z","lastTransitionTime":"2025-10-04T03:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.516252 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.516316 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.516334 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.516360 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.516383 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:52Z","lastTransitionTime":"2025-10-04T03:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.619434 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.619483 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.619495 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.619512 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.619526 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:52Z","lastTransitionTime":"2025-10-04T03:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.673123 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.673122 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:52 crc kubenswrapper[4770]: E1004 03:04:52.673628 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:52 crc kubenswrapper[4770]: E1004 03:04:52.673747 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.723659 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.723730 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.723747 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.723769 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.723787 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:52Z","lastTransitionTime":"2025-10-04T03:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.826906 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.826981 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.826999 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.827070 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.827094 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:52Z","lastTransitionTime":"2025-10-04T03:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.933071 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.933441 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.933533 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.933618 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:52 crc kubenswrapper[4770]: I1004 03:04:52.933703 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:52Z","lastTransitionTime":"2025-10-04T03:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.037480 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.037524 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.037536 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.037552 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.037565 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.141535 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.141586 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.141595 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.141628 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.141639 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.245404 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.245468 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.245484 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.245513 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.245530 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.348277 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.348360 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.348387 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.348428 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.348456 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.451981 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.452061 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.452075 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.452101 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.452117 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.556240 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.556286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.556298 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.556314 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.556325 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.659697 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.659770 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.659789 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.659815 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.659836 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.673583 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.673685 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:53 crc kubenswrapper[4770]: E1004 03:04:53.673790 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:53 crc kubenswrapper[4770]: E1004 03:04:53.674075 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.763964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.764087 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.764113 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.764143 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.764165 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.868316 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.868384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.868404 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.868432 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.868453 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.973160 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.973227 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.973245 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.973270 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:53 crc kubenswrapper[4770]: I1004 03:04:53.973289 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:53Z","lastTransitionTime":"2025-10-04T03:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.077550 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.077616 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.077636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.077662 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.077683 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:54Z","lastTransitionTime":"2025-10-04T03:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.181037 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.181113 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.181131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.181156 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.181173 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:54Z","lastTransitionTime":"2025-10-04T03:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.284286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.284342 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.284357 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.284379 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.284393 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:54Z","lastTransitionTime":"2025-10-04T03:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.387675 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.387754 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.387787 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.387817 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.387838 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:54Z","lastTransitionTime":"2025-10-04T03:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.491606 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.491673 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.491689 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.491712 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.491728 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:54Z","lastTransitionTime":"2025-10-04T03:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.596135 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.596232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.596255 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.596294 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.596319 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:54Z","lastTransitionTime":"2025-10-04T03:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.673765 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.673833 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:54 crc kubenswrapper[4770]: E1004 03:04:54.674088 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:54 crc kubenswrapper[4770]: E1004 03:04:54.674310 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.700064 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.700131 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.700148 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.700175 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.700192 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:54Z","lastTransitionTime":"2025-10-04T03:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.803787 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.803881 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.803904 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.803931 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.803949 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:54Z","lastTransitionTime":"2025-10-04T03:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.907911 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.907982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.908001 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.908071 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:54 crc kubenswrapper[4770]: I1004 03:04:54.908094 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:54Z","lastTransitionTime":"2025-10-04T03:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.011277 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.011338 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.011357 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.011384 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.011401 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.114864 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.114949 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.114976 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.115042 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.115070 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.218561 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.218620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.218633 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.218659 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.218674 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.321832 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.321905 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.321923 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.321950 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.321969 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.424985 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.425111 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.425140 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.425176 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.425200 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.528693 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.528790 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.528803 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.528829 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.528852 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.631982 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.632090 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.632113 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.632139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.632158 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.673116 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.673161 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:55 crc kubenswrapper[4770]: E1004 03:04:55.673397 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:55 crc kubenswrapper[4770]: E1004 03:04:55.673613 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.735658 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.735731 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.735750 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.735783 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.735804 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.839183 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.839451 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.839477 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.839514 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.839536 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.942552 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.942620 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.942638 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.942664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:55 crc kubenswrapper[4770]: I1004 03:04:55.942684 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:55Z","lastTransitionTime":"2025-10-04T03:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.046353 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.046416 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.046433 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.046459 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.046477 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.150204 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.150254 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.150265 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.150283 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.150298 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.254416 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.254501 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.254514 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.254556 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.254571 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.358667 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.358785 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.358814 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.358851 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.358880 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.462299 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.462376 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.462396 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.462426 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.462480 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.566447 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.566539 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.566558 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.566583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.566603 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.669609 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.669671 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.669687 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.669710 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.669727 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.673244 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.673424 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:56 crc kubenswrapper[4770]: E1004 03:04:56.673566 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:56 crc kubenswrapper[4770]: E1004 03:04:56.673682 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.772494 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.772548 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.772567 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.772591 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.772610 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.875460 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.875516 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.875532 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.875554 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.875572 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.979583 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.979646 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.979664 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.979691 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:56 crc kubenswrapper[4770]: I1004 03:04:56.979708 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:56Z","lastTransitionTime":"2025-10-04T03:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.083363 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.083435 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.083449 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.083475 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.083491 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:57Z","lastTransitionTime":"2025-10-04T03:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.189581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.189792 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.189825 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.189913 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.189948 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:57Z","lastTransitionTime":"2025-10-04T03:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.293448 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.293522 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.293540 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.293564 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.293580 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:57Z","lastTransitionTime":"2025-10-04T03:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.395878 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.395946 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.395964 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.395989 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.396030 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:57Z","lastTransitionTime":"2025-10-04T03:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.499215 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.499284 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.499304 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.499328 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.499349 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:57Z","lastTransitionTime":"2025-10-04T03:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.603535 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.603617 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.603640 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.603674 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.603702 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:57Z","lastTransitionTime":"2025-10-04T03:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.673676 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.673856 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:57 crc kubenswrapper[4770]: E1004 03:04:57.674186 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:04:57 crc kubenswrapper[4770]: E1004 03:04:57.674774 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.675418 4770 scope.go:117] "RemoveContainer" containerID="fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653" Oct 04 03:04:57 crc kubenswrapper[4770]: E1004 03:04:57.675760 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rtksw_openshift-ovn-kubernetes(61fd243e-61c3-4f86-8ee3-5c374a3e8ce2)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.706864 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.706934 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.706957 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.706986 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.707046 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:57Z","lastTransitionTime":"2025-10-04T03:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.809714 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.809767 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.809784 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.809810 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.809829 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:57Z","lastTransitionTime":"2025-10-04T03:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.913511 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.913562 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.913581 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.913606 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:57 crc kubenswrapper[4770]: I1004 03:04:57.913623 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:57Z","lastTransitionTime":"2025-10-04T03:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.017450 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.017528 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.017547 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.017575 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.017593 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.121520 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.121587 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.121608 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.121636 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.121653 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.225655 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.225723 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.225741 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.225768 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.225793 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.332075 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.332203 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.332229 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.332261 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.332296 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.436154 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.436242 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.436256 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.436280 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.436298 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.541456 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.541527 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.541545 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.541572 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.541594 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.645462 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.645519 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.645536 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.645560 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.645579 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.673534 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.673534 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:04:58 crc kubenswrapper[4770]: E1004 03:04:58.673737 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:04:58 crc kubenswrapper[4770]: E1004 03:04:58.673873 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.749139 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.749232 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.749253 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.749286 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.749308 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.852191 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.852258 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.852276 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.852301 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.852320 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.956187 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.956265 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.956299 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.956333 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:58 crc kubenswrapper[4770]: I1004 03:04:58.956355 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:58Z","lastTransitionTime":"2025-10-04T03:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.043723 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.043808 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.043829 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.043858 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.043882 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:59Z","lastTransitionTime":"2025-10-04T03:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.068649 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.068694 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.068704 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.068725 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.068737 4770 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:04:59Z","lastTransitionTime":"2025-10-04T03:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.103312 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt"] Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.103937 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.106484 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.107471 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.107495 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.108639 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.176445 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-z9xqp" podStartSLOduration=88.176418842 podStartE2EDuration="1m28.176418842s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:04:59.169344049 +0000 UTC m=+110.461353841" watchObservedRunningTime="2025-10-04 03:04:59.176418842 +0000 UTC m=+110.468428594" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.184228 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/21109bd0-78e0-44e0-9d8f-15e707b0d1df-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.184322 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21109bd0-78e0-44e0-9d8f-15e707b0d1df-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.184372 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/21109bd0-78e0-44e0-9d8f-15e707b0d1df-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.184448 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/21109bd0-78e0-44e0-9d8f-15e707b0d1df-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.184514 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21109bd0-78e0-44e0-9d8f-15e707b0d1df-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.285282 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/21109bd0-78e0-44e0-9d8f-15e707b0d1df-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.285367 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/21109bd0-78e0-44e0-9d8f-15e707b0d1df-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.285426 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21109bd0-78e0-44e0-9d8f-15e707b0d1df-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.285464 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/21109bd0-78e0-44e0-9d8f-15e707b0d1df-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.285517 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21109bd0-78e0-44e0-9d8f-15e707b0d1df-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.286096 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/21109bd0-78e0-44e0-9d8f-15e707b0d1df-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.286096 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/21109bd0-78e0-44e0-9d8f-15e707b0d1df-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.287208 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/21109bd0-78e0-44e0-9d8f-15e707b0d1df-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.295976 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21109bd0-78e0-44e0-9d8f-15e707b0d1df-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.302751 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21109bd0-78e0-44e0-9d8f-15e707b0d1df-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g75nt\" (UID: \"21109bd0-78e0-44e0-9d8f-15e707b0d1df\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.425794 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.673497 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:04:59 crc kubenswrapper[4770]: I1004 03:04:59.673513 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:04:59 crc kubenswrapper[4770]: E1004 03:04:59.674785 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:04:59 crc kubenswrapper[4770]: E1004 03:04:59.675043 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:00 crc kubenswrapper[4770]: I1004 03:05:00.400539 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" event={"ID":"21109bd0-78e0-44e0-9d8f-15e707b0d1df","Type":"ContainerStarted","Data":"b48316a3c23bec1814f164abae79f24051b4e924266be247bd899c9dc78ccde5"} Oct 04 03:05:00 crc kubenswrapper[4770]: I1004 03:05:00.400627 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" event={"ID":"21109bd0-78e0-44e0-9d8f-15e707b0d1df","Type":"ContainerStarted","Data":"95475099b87d7a4533bfdfab4945aa3d43ea591ac1e4ce99eafab6cad63ba488"} Oct 04 03:05:00 crc kubenswrapper[4770]: I1004 03:05:00.421991 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g75nt" podStartSLOduration=90.421887234 podStartE2EDuration="1m30.421887234s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:00.419761958 +0000 UTC m=+111.711771710" watchObservedRunningTime="2025-10-04 03:05:00.421887234 +0000 UTC m=+111.713896986" Oct 04 03:05:00 crc kubenswrapper[4770]: I1004 03:05:00.672940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:00 crc kubenswrapper[4770]: I1004 03:05:00.672940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:00 crc kubenswrapper[4770]: E1004 03:05:00.674144 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:00 crc kubenswrapper[4770]: E1004 03:05:00.674305 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:01 crc kubenswrapper[4770]: I1004 03:05:01.672853 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:01 crc kubenswrapper[4770]: I1004 03:05:01.672880 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:01 crc kubenswrapper[4770]: E1004 03:05:01.673020 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:01 crc kubenswrapper[4770]: E1004 03:05:01.673092 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:02 crc kubenswrapper[4770]: I1004 03:05:02.673530 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:02 crc kubenswrapper[4770]: E1004 03:05:02.673742 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:02 crc kubenswrapper[4770]: I1004 03:05:02.673535 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:02 crc kubenswrapper[4770]: E1004 03:05:02.674135 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:03 crc kubenswrapper[4770]: I1004 03:05:03.672908 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:03 crc kubenswrapper[4770]: I1004 03:05:03.672985 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:03 crc kubenswrapper[4770]: E1004 03:05:03.673121 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:03 crc kubenswrapper[4770]: E1004 03:05:03.673353 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:04 crc kubenswrapper[4770]: I1004 03:05:04.672892 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:04 crc kubenswrapper[4770]: E1004 03:05:04.673105 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:04 crc kubenswrapper[4770]: I1004 03:05:04.674059 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:04 crc kubenswrapper[4770]: E1004 03:05:04.674404 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:05 crc kubenswrapper[4770]: I1004 03:05:05.673671 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:05 crc kubenswrapper[4770]: I1004 03:05:05.673882 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:05 crc kubenswrapper[4770]: E1004 03:05:05.674065 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:05 crc kubenswrapper[4770]: E1004 03:05:05.674428 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:06 crc kubenswrapper[4770]: I1004 03:05:06.426896 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/1.log" Oct 04 03:05:06 crc kubenswrapper[4770]: I1004 03:05:06.428467 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/0.log" Oct 04 03:05:06 crc kubenswrapper[4770]: I1004 03:05:06.428545 4770 generic.go:334] "Generic (PLEG): container finished" podID="88d62119-65f3-4846-8b93-53d096fa9df7" containerID="6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0" exitCode=1 Oct 04 03:05:06 crc kubenswrapper[4770]: I1004 03:05:06.428589 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-66tn4" event={"ID":"88d62119-65f3-4846-8b93-53d096fa9df7","Type":"ContainerDied","Data":"6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0"} Oct 04 03:05:06 crc kubenswrapper[4770]: I1004 03:05:06.428639 4770 scope.go:117] "RemoveContainer" containerID="96cda5291b310e0127c3eb6b943cec389fd1219b9315315a630c038f7336544e" Oct 04 03:05:06 crc kubenswrapper[4770]: I1004 03:05:06.429602 4770 scope.go:117] "RemoveContainer" containerID="6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0" Oct 04 03:05:06 crc kubenswrapper[4770]: E1004 03:05:06.429925 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-66tn4_openshift-multus(88d62119-65f3-4846-8b93-53d096fa9df7)\"" pod="openshift-multus/multus-66tn4" podUID="88d62119-65f3-4846-8b93-53d096fa9df7" Oct 04 03:05:06 crc kubenswrapper[4770]: I1004 03:05:06.674595 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:06 crc kubenswrapper[4770]: E1004 03:05:06.675445 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:06 crc kubenswrapper[4770]: I1004 03:05:06.675574 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:06 crc kubenswrapper[4770]: E1004 03:05:06.675835 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:07 crc kubenswrapper[4770]: I1004 03:05:07.438627 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/1.log" Oct 04 03:05:07 crc kubenswrapper[4770]: I1004 03:05:07.672742 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:07 crc kubenswrapper[4770]: I1004 03:05:07.672947 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:07 crc kubenswrapper[4770]: E1004 03:05:07.673002 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:07 crc kubenswrapper[4770]: E1004 03:05:07.673969 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:08 crc kubenswrapper[4770]: I1004 03:05:08.673643 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:08 crc kubenswrapper[4770]: E1004 03:05:08.673847 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:08 crc kubenswrapper[4770]: I1004 03:05:08.673846 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:08 crc kubenswrapper[4770]: E1004 03:05:08.673991 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:09 crc kubenswrapper[4770]: E1004 03:05:09.619771 4770 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 04 03:05:09 crc kubenswrapper[4770]: I1004 03:05:09.673419 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:09 crc kubenswrapper[4770]: I1004 03:05:09.673604 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:09 crc kubenswrapper[4770]: E1004 03:05:09.680731 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:09 crc kubenswrapper[4770]: E1004 03:05:09.681417 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:09 crc kubenswrapper[4770]: E1004 03:05:09.825439 4770 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:05:10 crc kubenswrapper[4770]: I1004 03:05:10.672712 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:10 crc kubenswrapper[4770]: I1004 03:05:10.672819 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:10 crc kubenswrapper[4770]: E1004 03:05:10.672930 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:10 crc kubenswrapper[4770]: E1004 03:05:10.673224 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:11 crc kubenswrapper[4770]: I1004 03:05:11.672788 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:11 crc kubenswrapper[4770]: I1004 03:05:11.673093 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:11 crc kubenswrapper[4770]: E1004 03:05:11.673269 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:11 crc kubenswrapper[4770]: E1004 03:05:11.673973 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:11 crc kubenswrapper[4770]: I1004 03:05:11.674385 4770 scope.go:117] "RemoveContainer" containerID="fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653" Oct 04 03:05:12 crc kubenswrapper[4770]: I1004 03:05:12.464999 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/3.log" Oct 04 03:05:12 crc kubenswrapper[4770]: I1004 03:05:12.470352 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerStarted","Data":"80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50"} Oct 04 03:05:12 crc kubenswrapper[4770]: I1004 03:05:12.470905 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:05:12 crc kubenswrapper[4770]: I1004 03:05:12.504778 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podStartSLOduration=102.504757612 podStartE2EDuration="1m42.504757612s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:12.50158084 +0000 UTC m=+123.793590562" watchObservedRunningTime="2025-10-04 03:05:12.504757612 +0000 UTC m=+123.796767324" Oct 04 03:05:12 crc kubenswrapper[4770]: I1004 03:05:12.672795 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:12 crc kubenswrapper[4770]: I1004 03:05:12.672795 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:12 crc kubenswrapper[4770]: E1004 03:05:12.672941 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:12 crc kubenswrapper[4770]: E1004 03:05:12.673140 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:12 crc kubenswrapper[4770]: I1004 03:05:12.735263 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-krp9c"] Oct 04 03:05:12 crc kubenswrapper[4770]: I1004 03:05:12.735828 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:12 crc kubenswrapper[4770]: E1004 03:05:12.735991 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:13 crc kubenswrapper[4770]: I1004 03:05:13.673341 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:13 crc kubenswrapper[4770]: E1004 03:05:13.673619 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:14 crc kubenswrapper[4770]: I1004 03:05:14.672931 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:14 crc kubenswrapper[4770]: I1004 03:05:14.673123 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:14 crc kubenswrapper[4770]: I1004 03:05:14.673176 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:14 crc kubenswrapper[4770]: E1004 03:05:14.673383 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:14 crc kubenswrapper[4770]: E1004 03:05:14.673526 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:14 crc kubenswrapper[4770]: E1004 03:05:14.673721 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:14 crc kubenswrapper[4770]: E1004 03:05:14.826676 4770 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:05:15 crc kubenswrapper[4770]: I1004 03:05:15.673683 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:15 crc kubenswrapper[4770]: E1004 03:05:15.673956 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:16 crc kubenswrapper[4770]: I1004 03:05:16.673491 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:16 crc kubenswrapper[4770]: I1004 03:05:16.673540 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:16 crc kubenswrapper[4770]: I1004 03:05:16.673491 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:16 crc kubenswrapper[4770]: E1004 03:05:16.673769 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:16 crc kubenswrapper[4770]: E1004 03:05:16.673832 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:16 crc kubenswrapper[4770]: E1004 03:05:16.673919 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:17 crc kubenswrapper[4770]: I1004 03:05:17.672828 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:17 crc kubenswrapper[4770]: E1004 03:05:17.673122 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:18 crc kubenswrapper[4770]: I1004 03:05:18.673594 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:18 crc kubenswrapper[4770]: I1004 03:05:18.673734 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:18 crc kubenswrapper[4770]: I1004 03:05:18.673849 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:18 crc kubenswrapper[4770]: E1004 03:05:18.674080 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:18 crc kubenswrapper[4770]: E1004 03:05:18.674250 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:18 crc kubenswrapper[4770]: E1004 03:05:18.674355 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:19 crc kubenswrapper[4770]: I1004 03:05:19.673267 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:19 crc kubenswrapper[4770]: E1004 03:05:19.675719 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:19 crc kubenswrapper[4770]: E1004 03:05:19.827343 4770 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:05:20 crc kubenswrapper[4770]: I1004 03:05:20.673394 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:20 crc kubenswrapper[4770]: I1004 03:05:20.673394 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:20 crc kubenswrapper[4770]: I1004 03:05:20.673467 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:20 crc kubenswrapper[4770]: E1004 03:05:20.673881 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:20 crc kubenswrapper[4770]: E1004 03:05:20.673931 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:20 crc kubenswrapper[4770]: E1004 03:05:20.673999 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:20 crc kubenswrapper[4770]: I1004 03:05:20.674088 4770 scope.go:117] "RemoveContainer" containerID="6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0" Oct 04 03:05:21 crc kubenswrapper[4770]: I1004 03:05:21.511509 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/1.log" Oct 04 03:05:21 crc kubenswrapper[4770]: I1004 03:05:21.512119 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-66tn4" event={"ID":"88d62119-65f3-4846-8b93-53d096fa9df7","Type":"ContainerStarted","Data":"9f37453b28a502cbdaf06be847a0a05df4f86f3ee80d189151769d8da5ab3c91"} Oct 04 03:05:21 crc kubenswrapper[4770]: I1004 03:05:21.674054 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:21 crc kubenswrapper[4770]: E1004 03:05:21.674312 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:22 crc kubenswrapper[4770]: I1004 03:05:22.673674 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:22 crc kubenswrapper[4770]: E1004 03:05:22.673892 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:22 crc kubenswrapper[4770]: I1004 03:05:22.674809 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:22 crc kubenswrapper[4770]: I1004 03:05:22.674896 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:22 crc kubenswrapper[4770]: E1004 03:05:22.675259 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:22 crc kubenswrapper[4770]: E1004 03:05:22.675392 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:23 crc kubenswrapper[4770]: I1004 03:05:23.673535 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:23 crc kubenswrapper[4770]: E1004 03:05:23.673800 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:05:24 crc kubenswrapper[4770]: I1004 03:05:24.673243 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:24 crc kubenswrapper[4770]: I1004 03:05:24.673332 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:24 crc kubenswrapper[4770]: I1004 03:05:24.673357 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:24 crc kubenswrapper[4770]: E1004 03:05:24.673507 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:05:24 crc kubenswrapper[4770]: E1004 03:05:24.673609 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:05:24 crc kubenswrapper[4770]: E1004 03:05:24.673841 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krp9c" podUID="5e7fe8aa-4b62-47da-8aa7-0d31dcdba457" Oct 04 03:05:25 crc kubenswrapper[4770]: I1004 03:05:25.673273 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:25 crc kubenswrapper[4770]: I1004 03:05:25.676639 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 04 03:05:25 crc kubenswrapper[4770]: I1004 03:05:25.677864 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 04 03:05:26 crc kubenswrapper[4770]: I1004 03:05:26.672895 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:26 crc kubenswrapper[4770]: I1004 03:05:26.673094 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:26 crc kubenswrapper[4770]: I1004 03:05:26.673320 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:26 crc kubenswrapper[4770]: I1004 03:05:26.686895 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 04 03:05:26 crc kubenswrapper[4770]: I1004 03:05:26.686961 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 04 03:05:26 crc kubenswrapper[4770]: I1004 03:05:26.687081 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 04 03:05:26 crc kubenswrapper[4770]: I1004 03:05:26.687128 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.629190 4770 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.697143 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fbjnf"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.698220 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.698525 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.699068 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.699367 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.700166 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.726259 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.726259 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.727125 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.727283 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.727421 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.727648 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.727801 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.727992 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.728230 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.728466 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.728584 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xkkrp"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.728681 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.729740 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.730316 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.732982 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pnsm5"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.735814 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.736736 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.739122 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.744570 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.744652 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.744827 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.744954 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.744588 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.745230 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.745362 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.745466 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.745488 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ppsgj"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.745762 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.745977 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.746198 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.746237 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.746393 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.746644 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.746743 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.746984 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.748752 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.753934 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r8rk"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.754558 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.755092 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.755281 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.755675 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-xwhz4"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.756390 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.758367 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.758448 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.758685 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.759473 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-lwhtk"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.760197 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lwhtk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.761853 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.764766 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.764980 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.765231 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.765579 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.765696 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.765846 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.769780 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.770495 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.770533 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.770724 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.770862 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.771195 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.771449 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.772399 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.772569 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.772746 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.773169 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.773341 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.773452 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.773598 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.773722 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.773874 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.774066 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.774587 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.774731 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.774883 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.774999 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.775170 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pv4st"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.775301 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.775472 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x95pk"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.775498 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.775612 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.775748 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.775876 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.775943 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.775986 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.776116 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.776240 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.776463 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.777832 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.777931 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778071 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778117 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778211 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778076 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778416 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778448 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778486 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778548 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778565 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778643 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778664 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778742 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778820 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.778891 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.779174 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.779310 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.780719 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.781789 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.782185 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.782886 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.783676 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.783853 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.788056 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.788738 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.789045 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-kpt6s"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.789220 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.789492 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.789662 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.802370 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mwtcd"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.803198 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.804515 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.804836 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.805373 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.807854 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q7wr\" (UniqueName: \"kubernetes.io/projected/3b6ec72f-1166-4e03-815a-75d256d9ab5a-kube-api-access-5q7wr\") pod \"cluster-samples-operator-665b6dd947-n56bx\" (UID: \"3b6ec72f-1166-4e03-815a-75d256d9ab5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.807906 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3b6ec72f-1166-4e03-815a-75d256d9ab5a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-n56bx\" (UID: \"3b6ec72f-1166-4e03-815a-75d256d9ab5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.807958 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e1a8be1-36ca-4579-9678-13ba76cec6b1-config\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.807990 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5x52\" (UniqueName: \"kubernetes.io/projected/0e1a8be1-36ca-4579-9678-13ba76cec6b1-kube-api-access-l5x52\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808030 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-serving-cert\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808070 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-audit-policies\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808095 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e1a8be1-36ca-4579-9678-13ba76cec6b1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808115 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808160 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-encryption-config\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808197 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-audit-dir\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808309 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808338 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-etcd-client\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808364 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8f5s\" (UniqueName: \"kubernetes.io/projected/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-kube-api-access-q8f5s\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.808407 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e1a8be1-36ca-4579-9678-13ba76cec6b1-images\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.834363 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.834528 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.834617 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.834697 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.834799 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.834872 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.834977 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.834988 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.835107 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.835332 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.835507 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.835779 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.836502 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ggg8j"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.837470 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.837518 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.838051 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.839092 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.840136 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.840240 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.841100 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.841185 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.842117 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.846731 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.847195 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.847587 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.847609 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.849245 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-x2rjg"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.852453 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.854702 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.854988 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.856091 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.856382 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.857410 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.859304 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.863431 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.866912 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.869261 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.882096 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.883502 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.883509 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.885446 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.890106 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.890194 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lnhsd"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.896771 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.897140 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.898757 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.910616 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.912773 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.913839 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.914715 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915075 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q7wr\" (UniqueName: \"kubernetes.io/projected/3b6ec72f-1166-4e03-815a-75d256d9ab5a-kube-api-access-5q7wr\") pod \"cluster-samples-operator-665b6dd947-n56bx\" (UID: \"3b6ec72f-1166-4e03-815a-75d256d9ab5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915176 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9070d4f9-0fc3-4542-b010-b83193b54a0d-proxy-tls\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915252 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-trusted-ca-bundle\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915333 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915569 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfwfc\" (UniqueName: \"kubernetes.io/projected/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-kube-api-access-lfwfc\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915647 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3b6ec72f-1166-4e03-815a-75d256d9ab5a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-n56bx\" (UID: \"3b6ec72f-1166-4e03-815a-75d256d9ab5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915728 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b153ca4e-5d53-442d-8a73-72ff80ebbe1c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lt6hf\" (UID: \"b153ca4e-5d53-442d-8a73-72ff80ebbe1c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915803 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-service-ca\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915875 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.915952 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px668\" (UniqueName: \"kubernetes.io/projected/0a7275fe-266a-4920-a30d-28d3f0ee4a03-kube-api-access-px668\") pod \"dns-operator-744455d44c-x95pk\" (UID: \"0a7275fe-266a-4920-a30d-28d3f0ee4a03\") " pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916050 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a68c8d9f-b526-4583-901a-23595a0704a1-serving-cert\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916140 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42188dc5-d97c-423a-8562-4e7ece931916-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tmzsr\" (UID: \"42188dc5-d97c-423a-8562-4e7ece931916\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916231 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxc44\" (UniqueName: \"kubernetes.io/projected/61545ae7-a4d3-49dc-95a4-6393143a3718-kube-api-access-pxc44\") pod \"multus-admission-controller-857f4d67dd-kpt6s\" (UID: \"61545ae7-a4d3-49dc-95a4-6393143a3718\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916314 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a68c8d9f-b526-4583-901a-23595a0704a1-service-ca-bundle\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916403 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1483904-9014-4f9d-944b-3612fbd3b3d9-serving-cert\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916478 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-config\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916553 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916628 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a68c8d9f-b526-4583-901a-23595a0704a1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916704 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb105179-5c3e-4ec2-8204-6898e139c938-trusted-ca\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916777 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e1a8be1-36ca-4579-9678-13ba76cec6b1-config\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916851 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5x52\" (UniqueName: \"kubernetes.io/projected/0e1a8be1-36ca-4579-9678-13ba76cec6b1-kube-api-access-l5x52\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.916925 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61545ae7-a4d3-49dc-95a4-6393143a3718-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-kpt6s\" (UID: \"61545ae7-a4d3-49dc-95a4-6393143a3718\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917050 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44stl\" (UniqueName: \"kubernetes.io/projected/c7a98e59-558d-4731-a626-3dc946844eaa-kube-api-access-44stl\") pod \"openshift-controller-manager-operator-756b6f6bc6-nwxt6\" (UID: \"c7a98e59-558d-4731-a626-3dc946844eaa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917133 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-config\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917212 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917289 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/253c36e7-4eba-492f-8057-fd3865fcd223-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-h7lrz\" (UID: \"253c36e7-4eba-492f-8057-fd3865fcd223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917368 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-serving-cert\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917445 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-image-import-ca\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917515 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917592 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-client-ca\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917669 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941a4e7d-5fdb-4642-8d90-9a9cef86e912-serving-cert\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917757 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plt4z\" (UniqueName: \"kubernetes.io/projected/4bf7a322-09ed-493a-bd28-2639f15402f5-kube-api-access-plt4z\") pod \"downloads-7954f5f757-lwhtk\" (UID: \"4bf7a322-09ed-493a-bd28-2639f15402f5\") " pod="openshift-console/downloads-7954f5f757-lwhtk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917830 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-config\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917902 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68c8d9f-b526-4583-901a-23595a0704a1-config\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.917981 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-etcd-serving-ca\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.925878 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gczms\" (UniqueName: \"kubernetes.io/projected/45c9685e-2941-4778-97d2-f745955fa577-kube-api-access-gczms\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927039 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e5c5896d-a1f2-433c-858d-cea502c067d8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927317 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927419 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7a98e59-558d-4731-a626-3dc946844eaa-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nwxt6\" (UID: \"c7a98e59-558d-4731-a626-3dc946844eaa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927501 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-audit-policies\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927577 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-serving-cert\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927648 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/941a4e7d-5fdb-4642-8d90-9a9cef86e912-audit-dir\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927720 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh2bn\" (UniqueName: \"kubernetes.io/projected/a68c8d9f-b526-4583-901a-23595a0704a1-kube-api-access-mh2bn\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927796 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb105179-5c3e-4ec2-8204-6898e139c938-config\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927852 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3b6ec72f-1166-4e03-815a-75d256d9ab5a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-n56bx\" (UID: \"3b6ec72f-1166-4e03-815a-75d256d9ab5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.927878 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x29ns\" (UniqueName: \"kubernetes.io/projected/cb105179-5c3e-4ec2-8204-6898e139c938-kube-api-access-x29ns\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928032 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td2tv\" (UniqueName: \"kubernetes.io/projected/42188dc5-d97c-423a-8562-4e7ece931916-kube-api-access-td2tv\") pod \"openshift-apiserver-operator-796bbdcf4f-tmzsr\" (UID: \"42188dc5-d97c-423a-8562-4e7ece931916\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928126 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e1a8be1-36ca-4579-9678-13ba76cec6b1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928212 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928310 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928394 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928475 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bljqd\" (UniqueName: \"kubernetes.io/projected/e5c5896d-a1f2-433c-858d-cea502c067d8-kube-api-access-bljqd\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928559 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a7275fe-266a-4920-a30d-28d3f0ee4a03-metrics-tls\") pod \"dns-operator-744455d44c-x95pk\" (UID: \"0a7275fe-266a-4920-a30d-28d3f0ee4a03\") " pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928633 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjhlh\" (UniqueName: \"kubernetes.io/projected/2b45833c-2e11-428e-8131-1ea4676a794f-kube-api-access-vjhlh\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928700 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b1fc24e8-c0cc-499d-8498-bf591b24ee64-etcd-service-ca\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928770 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/253c36e7-4eba-492f-8057-fd3865fcd223-config\") pod \"kube-apiserver-operator-766d6c64bb-h7lrz\" (UID: \"253c36e7-4eba-492f-8057-fd3865fcd223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928841 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-config\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928916 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ec292e26-13cf-49b1-83a5-1c325bdd6d99-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.928993 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-encryption-config\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.929118 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb105179-5c3e-4ec2-8204-6898e139c938-serving-cert\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.929200 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/941a4e7d-5fdb-4642-8d90-9a9cef86e912-node-pullsecrets\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.929276 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm228\" (UniqueName: \"kubernetes.io/projected/941a4e7d-5fdb-4642-8d90-9a9cef86e912-kube-api-access-fm228\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.929361 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9070d4f9-0fc3-4542-b010-b83193b54a0d-images\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.929435 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42188dc5-d97c-423a-8562-4e7ece931916-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tmzsr\" (UID: \"42188dc5-d97c-423a-8562-4e7ece931916\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.929773 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvfs7\" (UniqueName: \"kubernetes.io/projected/c4c445d6-f62c-4d63-8d18-35441eebeb49-kube-api-access-vvfs7\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.929845 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1fc24e8-c0cc-499d-8498-bf591b24ee64-config\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.929931 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9070d4f9-0fc3-4542-b010-b83193b54a0d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930031 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/45c9685e-2941-4778-97d2-f745955fa577-audit-dir\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930110 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930185 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-audit-dir\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930259 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/941a4e7d-5fdb-4642-8d90-9a9cef86e912-etcd-client\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930331 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e5c5896d-a1f2-433c-858d-cea502c067d8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930408 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930474 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frncq\" (UniqueName: \"kubernetes.io/projected/a1483904-9014-4f9d-944b-3612fbd3b3d9-kube-api-access-frncq\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930542 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-auth-proxy-config\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930615 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c445d6-f62c-4d63-8d18-35441eebeb49-serving-cert\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930688 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930749 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/941a4e7d-5fdb-4642-8d90-9a9cef86e912-encryption-config\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930821 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-oauth-serving-cert\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930886 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a98e59-558d-4731-a626-3dc946844eaa-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nwxt6\" (UID: \"c7a98e59-558d-4731-a626-3dc946844eaa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.930989 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-etcd-client\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.931087 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b153ca4e-5d53-442d-8a73-72ff80ebbe1c-serving-cert\") pod \"openshift-config-operator-7777fb866f-lt6hf\" (UID: \"b153ca4e-5d53-442d-8a73-72ff80ebbe1c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.918941 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e1a8be1-36ca-4579-9678-13ba76cec6b1-config\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.931169 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ec292e26-13cf-49b1-83a5-1c325bdd6d99-metrics-tls\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.926485 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.931331 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.932299 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.926973 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.934552 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-serving-cert\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.935304 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-audit-dir\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.935805 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.936283 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937466 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-client-ca\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937686 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-audit\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937731 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8f5s\" (UniqueName: \"kubernetes.io/projected/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-kube-api-access-q8f5s\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937753 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw9lj\" (UniqueName: \"kubernetes.io/projected/9070d4f9-0fc3-4542-b010-b83193b54a0d-kube-api-access-pw9lj\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937809 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-oauth-config\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937890 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e1a8be1-36ca-4579-9678-13ba76cec6b1-images\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937928 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgnt9\" (UniqueName: \"kubernetes.io/projected/b153ca4e-5d53-442d-8a73-72ff80ebbe1c-kube-api-access-lgnt9\") pod \"openshift-config-operator-7777fb866f-lt6hf\" (UID: \"b153ca4e-5d53-442d-8a73-72ff80ebbe1c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937944 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1fc24e8-c0cc-499d-8498-bf591b24ee64-serving-cert\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937960 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkgvq\" (UniqueName: \"kubernetes.io/projected/ec292e26-13cf-49b1-83a5-1c325bdd6d99-kube-api-access-xkgvq\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.937978 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/253c36e7-4eba-492f-8057-fd3865fcd223-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-h7lrz\" (UID: \"253c36e7-4eba-492f-8057-fd3865fcd223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938021 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b1fc24e8-c0cc-499d-8498-bf591b24ee64-etcd-client\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938069 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938091 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-machine-approver-tls\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938110 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec292e26-13cf-49b1-83a5-1c325bdd6d99-trusted-ca\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938136 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-audit-policies\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938152 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938172 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5c5896d-a1f2-433c-858d-cea502c067d8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938193 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-console-config\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938210 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b1fc24e8-c0cc-499d-8498-bf591b24ee64-etcd-ca\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.938233 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-676nf\" (UniqueName: \"kubernetes.io/projected/b1fc24e8-c0cc-499d-8498-bf591b24ee64-kube-api-access-676nf\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.939142 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e1a8be1-36ca-4579-9678-13ba76cec6b1-images\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.940402 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ppsgj"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.940659 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-audit-policies\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.941985 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.942750 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-encryption-config\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.942825 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.942838 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r8rk"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.957554 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2fqvm"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.944136 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.944413 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e1a8be1-36ca-4579-9678-13ba76cec6b1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.950040 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-etcd-client\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.959543 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jt2cp"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.959707 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.962868 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.963150 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.963361 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.964322 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fbjnf"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.964355 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pnsm5"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.964456 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.968540 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.970459 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-kpt6s"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.971599 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.974214 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x95pk"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.975795 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xwhz4"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.977546 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xkkrp"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.979476 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.981674 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lwhtk"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.983045 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.985437 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pv4st"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.986564 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.986912 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.988431 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.990176 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.991439 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.992425 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.994342 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vxv67"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.995406 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.995460 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-twts6"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.996439 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-twts6" Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.996505 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph"] Oct 04 03:05:29 crc kubenswrapper[4770]: I1004 03:05:29.999800 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mwtcd"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.002848 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.004739 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.007563 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.007587 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.011214 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x2rjg"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.015376 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.018569 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.021108 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.021483 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lnhsd"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.023473 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.026475 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-twts6"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.029590 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2fqvm"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.032197 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.033743 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.035340 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vxv67"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.037608 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-v57fn"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.039066 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.039336 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jt2cp"] Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040372 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-console-config\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040416 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b1fc24e8-c0cc-499d-8498-bf591b24ee64-etcd-ca\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040446 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-676nf\" (UniqueName: \"kubernetes.io/projected/b1fc24e8-c0cc-499d-8498-bf591b24ee64-kube-api-access-676nf\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040473 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5c5896d-a1f2-433c-858d-cea502c067d8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040515 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9070d4f9-0fc3-4542-b010-b83193b54a0d-proxy-tls\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040539 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-trusted-ca-bundle\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040564 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040599 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b153ca4e-5d53-442d-8a73-72ff80ebbe1c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lt6hf\" (UID: \"b153ca4e-5d53-442d-8a73-72ff80ebbe1c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040623 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-service-ca\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040653 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040679 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040681 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040854 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfwfc\" (UniqueName: \"kubernetes.io/projected/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-kube-api-access-lfwfc\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040879 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42188dc5-d97c-423a-8562-4e7ece931916-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tmzsr\" (UID: \"42188dc5-d97c-423a-8562-4e7ece931916\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040902 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px668\" (UniqueName: \"kubernetes.io/projected/0a7275fe-266a-4920-a30d-28d3f0ee4a03-kube-api-access-px668\") pod \"dns-operator-744455d44c-x95pk\" (UID: \"0a7275fe-266a-4920-a30d-28d3f0ee4a03\") " pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040920 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a68c8d9f-b526-4583-901a-23595a0704a1-serving-cert\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040939 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxc44\" (UniqueName: \"kubernetes.io/projected/61545ae7-a4d3-49dc-95a4-6393143a3718-kube-api-access-pxc44\") pod \"multus-admission-controller-857f4d67dd-kpt6s\" (UID: \"61545ae7-a4d3-49dc-95a4-6393143a3718\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040960 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a68c8d9f-b526-4583-901a-23595a0704a1-service-ca-bundle\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040978 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1483904-9014-4f9d-944b-3612fbd3b3d9-serving-cert\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.040996 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041033 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-config\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041068 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61545ae7-a4d3-49dc-95a4-6393143a3718-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-kpt6s\" (UID: \"61545ae7-a4d3-49dc-95a4-6393143a3718\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041085 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44stl\" (UniqueName: \"kubernetes.io/projected/c7a98e59-558d-4731-a626-3dc946844eaa-kube-api-access-44stl\") pod \"openshift-controller-manager-operator-756b6f6bc6-nwxt6\" (UID: \"c7a98e59-558d-4731-a626-3dc946844eaa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041106 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a68c8d9f-b526-4583-901a-23595a0704a1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041126 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb105179-5c3e-4ec2-8204-6898e139c938-trusted-ca\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041148 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-config\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041169 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-image-import-ca\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041187 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041205 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041223 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/253c36e7-4eba-492f-8057-fd3865fcd223-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-h7lrz\" (UID: \"253c36e7-4eba-492f-8057-fd3865fcd223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041243 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-client-ca\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041260 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941a4e7d-5fdb-4642-8d90-9a9cef86e912-serving-cert\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041288 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plt4z\" (UniqueName: \"kubernetes.io/projected/4bf7a322-09ed-493a-bd28-2639f15402f5-kube-api-access-plt4z\") pod \"downloads-7954f5f757-lwhtk\" (UID: \"4bf7a322-09ed-493a-bd28-2639f15402f5\") " pod="openshift-console/downloads-7954f5f757-lwhtk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041307 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-etcd-serving-ca\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041324 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gczms\" (UniqueName: \"kubernetes.io/projected/45c9685e-2941-4778-97d2-f745955fa577-kube-api-access-gczms\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041343 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-config\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041360 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68c8d9f-b526-4583-901a-23595a0704a1-config\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041377 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-console-config\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041379 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e5c5896d-a1f2-433c-858d-cea502c067d8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041424 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-serving-cert\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041446 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/941a4e7d-5fdb-4642-8d90-9a9cef86e912-audit-dir\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041466 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041483 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7a98e59-558d-4731-a626-3dc946844eaa-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nwxt6\" (UID: \"c7a98e59-558d-4731-a626-3dc946844eaa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041503 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x29ns\" (UniqueName: \"kubernetes.io/projected/cb105179-5c3e-4ec2-8204-6898e139c938-kube-api-access-x29ns\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041521 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh2bn\" (UniqueName: \"kubernetes.io/projected/a68c8d9f-b526-4583-901a-23595a0704a1-kube-api-access-mh2bn\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041539 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb105179-5c3e-4ec2-8204-6898e139c938-config\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041559 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td2tv\" (UniqueName: \"kubernetes.io/projected/42188dc5-d97c-423a-8562-4e7ece931916-kube-api-access-td2tv\") pod \"openshift-apiserver-operator-796bbdcf4f-tmzsr\" (UID: \"42188dc5-d97c-423a-8562-4e7ece931916\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041577 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041594 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041613 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bljqd\" (UniqueName: \"kubernetes.io/projected/e5c5896d-a1f2-433c-858d-cea502c067d8-kube-api-access-bljqd\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041663 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a7275fe-266a-4920-a30d-28d3f0ee4a03-metrics-tls\") pod \"dns-operator-744455d44c-x95pk\" (UID: \"0a7275fe-266a-4920-a30d-28d3f0ee4a03\") " pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041698 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjhlh\" (UniqueName: \"kubernetes.io/projected/2b45833c-2e11-428e-8131-1ea4676a794f-kube-api-access-vjhlh\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041751 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b1fc24e8-c0cc-499d-8498-bf591b24ee64-etcd-service-ca\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041767 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/253c36e7-4eba-492f-8057-fd3865fcd223-config\") pod \"kube-apiserver-operator-766d6c64bb-h7lrz\" (UID: \"253c36e7-4eba-492f-8057-fd3865fcd223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041799 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-config\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041817 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ec292e26-13cf-49b1-83a5-1c325bdd6d99-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.041833 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb105179-5c3e-4ec2-8204-6898e139c938-serving-cert\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042315 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9070d4f9-0fc3-4542-b010-b83193b54a0d-images\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042342 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/941a4e7d-5fdb-4642-8d90-9a9cef86e912-node-pullsecrets\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042358 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e5c5896d-a1f2-433c-858d-cea502c067d8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042360 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm228\" (UniqueName: \"kubernetes.io/projected/941a4e7d-5fdb-4642-8d90-9a9cef86e912-kube-api-access-fm228\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042399 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42188dc5-d97c-423a-8562-4e7ece931916-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tmzsr\" (UID: \"42188dc5-d97c-423a-8562-4e7ece931916\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042420 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9070d4f9-0fc3-4542-b010-b83193b54a0d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042438 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvfs7\" (UniqueName: \"kubernetes.io/projected/c4c445d6-f62c-4d63-8d18-35441eebeb49-kube-api-access-vvfs7\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042455 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1fc24e8-c0cc-499d-8498-bf591b24ee64-config\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042481 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/45c9685e-2941-4778-97d2-f745955fa577-audit-dir\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042497 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042519 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frncq\" (UniqueName: \"kubernetes.io/projected/a1483904-9014-4f9d-944b-3612fbd3b3d9-kube-api-access-frncq\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042535 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/941a4e7d-5fdb-4642-8d90-9a9cef86e912-etcd-client\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042551 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e5c5896d-a1f2-433c-858d-cea502c067d8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042573 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c445d6-f62c-4d63-8d18-35441eebeb49-serving-cert\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042589 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/941a4e7d-5fdb-4642-8d90-9a9cef86e912-encryption-config\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042623 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-auth-proxy-config\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042641 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-oauth-serving-cert\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042657 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a98e59-558d-4731-a626-3dc946844eaa-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nwxt6\" (UID: \"c7a98e59-558d-4731-a626-3dc946844eaa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042682 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b153ca4e-5d53-442d-8a73-72ff80ebbe1c-serving-cert\") pod \"openshift-config-operator-7777fb866f-lt6hf\" (UID: \"b153ca4e-5d53-442d-8a73-72ff80ebbe1c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042701 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ec292e26-13cf-49b1-83a5-1c325bdd6d99-metrics-tls\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042728 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw9lj\" (UniqueName: \"kubernetes.io/projected/9070d4f9-0fc3-4542-b010-b83193b54a0d-kube-api-access-pw9lj\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042746 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-oauth-config\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042763 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-client-ca\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042783 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-audit\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042801 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1fc24e8-c0cc-499d-8498-bf591b24ee64-serving-cert\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042820 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkgvq\" (UniqueName: \"kubernetes.io/projected/ec292e26-13cf-49b1-83a5-1c325bdd6d99-kube-api-access-xkgvq\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042835 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/253c36e7-4eba-492f-8057-fd3865fcd223-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-h7lrz\" (UID: \"253c36e7-4eba-492f-8057-fd3865fcd223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042864 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgnt9\" (UniqueName: \"kubernetes.io/projected/b153ca4e-5d53-442d-8a73-72ff80ebbe1c-kube-api-access-lgnt9\") pod \"openshift-config-operator-7777fb866f-lt6hf\" (UID: \"b153ca4e-5d53-442d-8a73-72ff80ebbe1c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042882 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b1fc24e8-c0cc-499d-8498-bf591b24ee64-etcd-client\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042900 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042918 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-machine-approver-tls\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042935 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec292e26-13cf-49b1-83a5-1c325bdd6d99-trusted-ca\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042951 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-audit-policies\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.042969 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.043184 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/941a4e7d-5fdb-4642-8d90-9a9cef86e912-audit-dir\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.043576 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-service-ca\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.043892 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.044289 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-trusted-ca-bundle\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.044780 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.045109 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b153ca4e-5d53-442d-8a73-72ff80ebbe1c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lt6hf\" (UID: \"b153ca4e-5d53-442d-8a73-72ff80ebbe1c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.045453 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.045478 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/45c9685e-2941-4778-97d2-f745955fa577-audit-dir\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.045503 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a68c8d9f-b526-4583-901a-23595a0704a1-service-ca-bundle\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.046101 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb105179-5c3e-4ec2-8204-6898e139c938-config\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.046889 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.047019 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.047398 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-config\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.047716 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.047838 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5c5896d-a1f2-433c-858d-cea502c067d8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.048092 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a68c8d9f-b526-4583-901a-23595a0704a1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.048367 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb105179-5c3e-4ec2-8204-6898e139c938-trusted-ca\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.048845 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ec292e26-13cf-49b1-83a5-1c325bdd6d99-metrics-tls\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.048976 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-config\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.049041 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-etcd-serving-ca\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.049368 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ec292e26-13cf-49b1-83a5-1c325bdd6d99-trusted-ca\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.049399 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-audit-policies\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.049756 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.049857 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.049919 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-oauth-config\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.050351 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7a98e59-558d-4731-a626-3dc946844eaa-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nwxt6\" (UID: \"c7a98e59-558d-4731-a626-3dc946844eaa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.050497 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a68c8d9f-b526-4583-901a-23595a0704a1-config\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.050747 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-config\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.050856 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42188dc5-d97c-423a-8562-4e7ece931916-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tmzsr\" (UID: \"42188dc5-d97c-423a-8562-4e7ece931916\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.051742 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-image-import-ca\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.051825 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-machine-approver-tls\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.052044 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.052046 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.052131 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.052847 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/941a4e7d-5fdb-4642-8d90-9a9cef86e912-node-pullsecrets\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.052940 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-client-ca\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.053116 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-config\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.053175 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/941a4e7d-5fdb-4642-8d90-9a9cef86e912-encryption-config\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.053518 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/941a4e7d-5fdb-4642-8d90-9a9cef86e912-etcd-client\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.053818 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a98e59-558d-4731-a626-3dc946844eaa-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nwxt6\" (UID: \"c7a98e59-558d-4731-a626-3dc946844eaa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.054061 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-oauth-serving-cert\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.054150 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1483904-9014-4f9d-944b-3612fbd3b3d9-serving-cert\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.054390 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42188dc5-d97c-423a-8562-4e7ece931916-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tmzsr\" (UID: \"42188dc5-d97c-423a-8562-4e7ece931916\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.054401 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/941a4e7d-5fdb-4642-8d90-9a9cef86e912-audit\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.054648 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.054947 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-client-ca\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.054968 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a7275fe-266a-4920-a30d-28d3f0ee4a03-metrics-tls\") pod \"dns-operator-744455d44c-x95pk\" (UID: \"0a7275fe-266a-4920-a30d-28d3f0ee4a03\") " pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.054997 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-auth-proxy-config\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.055438 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9070d4f9-0fc3-4542-b010-b83193b54a0d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.055860 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c445d6-f62c-4d63-8d18-35441eebeb49-serving-cert\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.056182 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941a4e7d-5fdb-4642-8d90-9a9cef86e912-serving-cert\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.056652 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-serving-cert\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.057522 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb105179-5c3e-4ec2-8204-6898e139c938-serving-cert\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.058684 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.059350 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b153ca4e-5d53-442d-8a73-72ff80ebbe1c-serving-cert\") pod \"openshift-config-operator-7777fb866f-lt6hf\" (UID: \"b153ca4e-5d53-442d-8a73-72ff80ebbe1c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.059924 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a68c8d9f-b526-4583-901a-23595a0704a1-serving-cert\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.060713 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.081370 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.091090 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/61545ae7-a4d3-49dc-95a4-6393143a3718-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-kpt6s\" (UID: \"61545ae7-a4d3-49dc-95a4-6393143a3718\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.101754 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.121169 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.141777 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.145850 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/253c36e7-4eba-492f-8057-fd3865fcd223-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-h7lrz\" (UID: \"253c36e7-4eba-492f-8057-fd3865fcd223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.161405 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.172177 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/253c36e7-4eba-492f-8057-fd3865fcd223-config\") pod \"kube-apiserver-operator-766d6c64bb-h7lrz\" (UID: \"253c36e7-4eba-492f-8057-fd3865fcd223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.181105 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.181837 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9070d4f9-0fc3-4542-b010-b83193b54a0d-images\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.201488 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.212614 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b1fc24e8-c0cc-499d-8498-bf591b24ee64-etcd-client\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.221935 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.240937 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.246969 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9070d4f9-0fc3-4542-b010-b83193b54a0d-proxy-tls\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.262148 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.281133 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.288245 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1fc24e8-c0cc-499d-8498-bf591b24ee64-serving-cert\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.301770 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.303499 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b1fc24e8-c0cc-499d-8498-bf591b24ee64-etcd-ca\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.321308 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.325525 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1fc24e8-c0cc-499d-8498-bf591b24ee64-config\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.342299 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.362118 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.381190 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.391741 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b1fc24e8-c0cc-499d-8498-bf591b24ee64-etcd-service-ca\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.401246 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.420912 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.441797 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.461874 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.481722 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.501840 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.522679 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.541406 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.560958 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.581485 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.602376 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.623306 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.642414 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.662082 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.681177 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.701342 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.723532 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.742437 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.762041 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.781066 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.801995 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.822057 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.840761 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.859658 4770 request.go:700] Waited for 1.011447459s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0 Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.862952 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.881671 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.901955 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.921432 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.943169 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.961715 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 04 03:05:30 crc kubenswrapper[4770]: I1004 03:05:30.983741 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.001520 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.022713 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.042197 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.061320 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.082775 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.100522 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.122686 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.141993 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.161872 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.181647 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.233720 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q7wr\" (UniqueName: \"kubernetes.io/projected/3b6ec72f-1166-4e03-815a-75d256d9ab5a-kube-api-access-5q7wr\") pod \"cluster-samples-operator-665b6dd947-n56bx\" (UID: \"3b6ec72f-1166-4e03-815a-75d256d9ab5a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.242193 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.254584 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5x52\" (UniqueName: \"kubernetes.io/projected/0e1a8be1-36ca-4579-9678-13ba76cec6b1-kube-api-access-l5x52\") pod \"machine-api-operator-5694c8668f-fbjnf\" (UID: \"0e1a8be1-36ca-4579-9678-13ba76cec6b1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.261292 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.270577 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.308949 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8f5s\" (UniqueName: \"kubernetes.io/projected/d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7-kube-api-access-q8f5s\") pod \"apiserver-7bbb656c7d-5m7kk\" (UID: \"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.321947 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.341575 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.362207 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.381484 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.410478 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.422557 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.442137 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.461686 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.482694 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.503465 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.520437 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.527832 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx"] Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.541949 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.554510 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.561820 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.575930 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.582348 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.601473 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.622459 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.644113 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.662284 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.681296 4770 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.702640 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.721247 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.741082 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.761331 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.785465 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk"] Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.802999 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.804579 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fbjnf"] Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.820717 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.840687 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.879163 4770 request.go:700] Waited for 1.836706713s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver/serviceaccounts/openshift-apiserver-sa/token Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.881733 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-676nf\" (UniqueName: \"kubernetes.io/projected/b1fc24e8-c0cc-499d-8498-bf591b24ee64-kube-api-access-676nf\") pod \"etcd-operator-b45778765-mwtcd\" (UID: \"b1fc24e8-c0cc-499d-8498-bf591b24ee64\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.903722 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm228\" (UniqueName: \"kubernetes.io/projected/941a4e7d-5fdb-4642-8d90-9a9cef86e912-kube-api-access-fm228\") pod \"apiserver-76f77b778f-pnsm5\" (UID: \"941a4e7d-5fdb-4642-8d90-9a9cef86e912\") " pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.920167 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfwfc\" (UniqueName: \"kubernetes.io/projected/2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4-kube-api-access-lfwfc\") pod \"machine-approver-56656f9798-ml6vl\" (UID: \"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.937780 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px668\" (UniqueName: \"kubernetes.io/projected/0a7275fe-266a-4920-a30d-28d3f0ee4a03-kube-api-access-px668\") pod \"dns-operator-744455d44c-x95pk\" (UID: \"0a7275fe-266a-4920-a30d-28d3f0ee4a03\") " pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.941554 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.964288 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x29ns\" (UniqueName: \"kubernetes.io/projected/cb105179-5c3e-4ec2-8204-6898e139c938-kube-api-access-x29ns\") pod \"console-operator-58897d9998-pv4st\" (UID: \"cb105179-5c3e-4ec2-8204-6898e139c938\") " pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:31 crc kubenswrapper[4770]: I1004 03:05:31.981711 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxc44\" (UniqueName: \"kubernetes.io/projected/61545ae7-a4d3-49dc-95a4-6393143a3718-kube-api-access-pxc44\") pod \"multus-admission-controller-857f4d67dd-kpt6s\" (UID: \"61545ae7-a4d3-49dc-95a4-6393143a3718\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.012348 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.026437 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e5c5896d-a1f2-433c-858d-cea502c067d8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.045684 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw9lj\" (UniqueName: \"kubernetes.io/projected/9070d4f9-0fc3-4542-b010-b83193b54a0d-kube-api-access-pw9lj\") pod \"machine-config-operator-74547568cd-gzgxr\" (UID: \"9070d4f9-0fc3-4542-b010-b83193b54a0d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.063657 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh2bn\" (UniqueName: \"kubernetes.io/projected/a68c8d9f-b526-4583-901a-23595a0704a1-kube-api-access-mh2bn\") pod \"authentication-operator-69f744f599-ppsgj\" (UID: \"a68c8d9f-b526-4583-901a-23595a0704a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.063791 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44stl\" (UniqueName: \"kubernetes.io/projected/c7a98e59-558d-4731-a626-3dc946844eaa-kube-api-access-44stl\") pod \"openshift-controller-manager-operator-756b6f6bc6-nwxt6\" (UID: \"c7a98e59-558d-4731-a626-3dc946844eaa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.082561 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.090567 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.091327 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgnt9\" (UniqueName: \"kubernetes.io/projected/b153ca4e-5d53-442d-8a73-72ff80ebbe1c-kube-api-access-lgnt9\") pod \"openshift-config-operator-7777fb866f-lt6hf\" (UID: \"b153ca4e-5d53-442d-8a73-72ff80ebbe1c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.095680 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gczms\" (UniqueName: \"kubernetes.io/projected/45c9685e-2941-4778-97d2-f745955fa577-kube-api-access-gczms\") pod \"oauth-openshift-558db77b4-6r8rk\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.112472 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.124131 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frncq\" (UniqueName: \"kubernetes.io/projected/a1483904-9014-4f9d-944b-3612fbd3b3d9-kube-api-access-frncq\") pod \"controller-manager-879f6c89f-xkkrp\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.138846 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.139083 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pnsm5"] Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.141297 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvfs7\" (UniqueName: \"kubernetes.io/projected/c4c445d6-f62c-4d63-8d18-35441eebeb49-kube-api-access-vvfs7\") pod \"route-controller-manager-6576b87f9c-jq7w9\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.162327 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ec292e26-13cf-49b1-83a5-1c325bdd6d99-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:32 crc kubenswrapper[4770]: W1004 03:05:32.164271 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod941a4e7d_5fdb_4642_8d90_9a9cef86e912.slice/crio-c0bf10788721118367adcd8840a074718ef88ea8267e92802d4b9cbc736f7070 WatchSource:0}: Error finding container c0bf10788721118367adcd8840a074718ef88ea8267e92802d4b9cbc736f7070: Status 404 returned error can't find the container with id c0bf10788721118367adcd8840a074718ef88ea8267e92802d4b9cbc736f7070 Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.173867 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.181504 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.188154 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td2tv\" (UniqueName: \"kubernetes.io/projected/42188dc5-d97c-423a-8562-4e7ece931916-kube-api-access-td2tv\") pod \"openshift-apiserver-operator-796bbdcf4f-tmzsr\" (UID: \"42188dc5-d97c-423a-8562-4e7ece931916\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.198446 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bljqd\" (UniqueName: \"kubernetes.io/projected/e5c5896d-a1f2-433c-858d-cea502c067d8-kube-api-access-bljqd\") pod \"cluster-image-registry-operator-dc59b4c8b-fbfg2\" (UID: \"e5c5896d-a1f2-433c-858d-cea502c067d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.203614 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.227513 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjhlh\" (UniqueName: \"kubernetes.io/projected/2b45833c-2e11-428e-8131-1ea4676a794f-kube-api-access-vjhlh\") pod \"console-f9d7485db-xwhz4\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.254369 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.270908 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plt4z\" (UniqueName: \"kubernetes.io/projected/4bf7a322-09ed-493a-bd28-2639f15402f5-kube-api-access-plt4z\") pod \"downloads-7954f5f757-lwhtk\" (UID: \"4bf7a322-09ed-493a-bd28-2639f15402f5\") " pod="openshift-console/downloads-7954f5f757-lwhtk" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.272890 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkgvq\" (UniqueName: \"kubernetes.io/projected/ec292e26-13cf-49b1-83a5-1c325bdd6d99-kube-api-access-xkgvq\") pod \"ingress-operator-5b745b69d9-zvgls\" (UID: \"ec292e26-13cf-49b1-83a5-1c325bdd6d99\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.282331 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.284707 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.293555 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.328639 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.328648 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pv4st"] Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.336731 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/253c36e7-4eba-492f-8057-fd3865fcd223-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-h7lrz\" (UID: \"253c36e7-4eba-492f-8057-fd3865fcd223\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:32 crc kubenswrapper[4770]: W1004 03:05:32.358788 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb105179_5c3e_4ec2_8204_6898e139c938.slice/crio-9a385081841c51576a5676d833094f2c71039371d50dd58f824da57ce8245e03 WatchSource:0}: Error finding container 9a385081841c51576a5676d833094f2c71039371d50dd58f824da57ce8245e03: Status 404 returned error can't find the container with id 9a385081841c51576a5676d833094f2c71039371d50dd58f824da57ce8245e03 Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.367916 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.375508 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lwhtk" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.385988 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-registry-certificates\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386196 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a5443b52-3d7f-4ff3-9c01-163da49074e2-tmpfs\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386260 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f922cbae-d663-4a41-96ed-fe20b4bc3abb-profile-collector-cert\") pod \"catalog-operator-68c6474976-s26vf\" (UID: \"f922cbae-d663-4a41-96ed-fe20b4bc3abb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386276 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2gh7\" (UniqueName: \"kubernetes.io/projected/f922cbae-d663-4a41-96ed-fe20b4bc3abb-kube-api-access-w2gh7\") pod \"catalog-operator-68c6474976-s26vf\" (UID: \"f922cbae-d663-4a41-96ed-fe20b4bc3abb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386388 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14ca55a-159f-48e5-a290-5462a4f8811c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fxv82\" (UID: \"e14ca55a-159f-48e5-a290-5462a4f8811c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386444 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2g8l\" (UniqueName: \"kubernetes.io/projected/03505445-42fb-4015-97d4-6d41167cadd1-kube-api-access-b2g8l\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386558 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5443b52-3d7f-4ff3-9c01-163da49074e2-webhook-cert\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386606 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e14ca55a-159f-48e5-a290-5462a4f8811c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fxv82\" (UID: \"e14ca55a-159f-48e5-a290-5462a4f8811c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386622 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ghlx\" (UniqueName: \"kubernetes.io/projected/e14ca55a-159f-48e5-a290-5462a4f8811c-kube-api-access-2ghlx\") pod \"kube-storage-version-migrator-operator-b67b599dd-fxv82\" (UID: \"e14ca55a-159f-48e5-a290-5462a4f8811c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386664 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5443b52-3d7f-4ff3-9c01-163da49074e2-apiservice-cert\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386703 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03505445-42fb-4015-97d4-6d41167cadd1-metrics-certs\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386718 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/108c220a-bc76-4e17-884d-80f2a5b63357-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386744 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cztnz\" (UniqueName: \"kubernetes.io/projected/e111a15d-a0c1-470b-9723-e4bbcc2265c1-kube-api-access-cztnz\") pod \"collect-profiles-29325780-z72ph\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386774 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03505445-42fb-4015-97d4-6d41167cadd1-service-ca-bundle\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386793 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3e950818-24fa-4cbe-bdd9-1dbb895b12f1-metrics-tls\") pod \"dns-default-x2rjg\" (UID: \"3e950818-24fa-4cbe-bdd9-1dbb895b12f1\") " pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386826 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386861 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e111a15d-a0c1-470b-9723-e4bbcc2265c1-secret-volume\") pod \"collect-profiles-29325780-z72ph\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386877 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0035edd0-6338-48f1-9534-3b80e135f956-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rgdpf\" (UID: \"0035edd0-6338-48f1-9534-3b80e135f956\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386893 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxsv9\" (UniqueName: \"kubernetes.io/projected/a5443b52-3d7f-4ff3-9c01-163da49074e2-kube-api-access-sxsv9\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.386989 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9dmn\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-kube-api-access-j9dmn\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387043 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs84j\" (UniqueName: \"kubernetes.io/projected/0035edd0-6338-48f1-9534-3b80e135f956-kube-api-access-rs84j\") pod \"machine-config-controller-84d6567774-rgdpf\" (UID: \"0035edd0-6338-48f1-9534-3b80e135f956\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387087 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/477ecb7c-f74b-4f17-87c6-e25d129aaacb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qq8rm\" (UID: \"477ecb7c-f74b-4f17-87c6-e25d129aaacb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387103 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g8rqs\" (UID: \"7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387200 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d25eff-feac-415e-9ba2-e34de3c0d866-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nmsxt\" (UID: \"e9d25eff-feac-415e-9ba2-e34de3c0d866\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387238 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d25eff-feac-415e-9ba2-e34de3c0d866-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nmsxt\" (UID: \"e9d25eff-feac-415e-9ba2-e34de3c0d866\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387285 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/056a8276-f8ee-45f5-a145-63db980dc0c4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8srq6\" (UID: \"056a8276-f8ee-45f5-a145-63db980dc0c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387311 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29mlx\" (UniqueName: \"kubernetes.io/projected/f54b9ac6-dfe8-4502-95c7-f657e8be4b05-kube-api-access-29mlx\") pod \"migrator-59844c95c7-bzk7x\" (UID: \"f54b9ac6-dfe8-4502-95c7-f657e8be4b05\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387332 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d25eff-feac-415e-9ba2-e34de3c0d866-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nmsxt\" (UID: \"e9d25eff-feac-415e-9ba2-e34de3c0d866\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387374 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/03505445-42fb-4015-97d4-6d41167cadd1-stats-auth\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387389 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f922cbae-d663-4a41-96ed-fe20b4bc3abb-srv-cert\") pod \"catalog-operator-68c6474976-s26vf\" (UID: \"f922cbae-d663-4a41-96ed-fe20b4bc3abb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387404 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e950818-24fa-4cbe-bdd9-1dbb895b12f1-config-volume\") pod \"dns-default-x2rjg\" (UID: \"3e950818-24fa-4cbe-bdd9-1dbb895b12f1\") " pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387449 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-bound-sa-token\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387485 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-registry-tls\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387500 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/108c220a-bc76-4e17-884d-80f2a5b63357-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387541 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0035edd0-6338-48f1-9534-3b80e135f956-proxy-tls\") pod \"machine-config-controller-84d6567774-rgdpf\" (UID: \"0035edd0-6338-48f1-9534-3b80e135f956\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387578 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e111a15d-a0c1-470b-9723-e4bbcc2265c1-config-volume\") pod \"collect-profiles-29325780-z72ph\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387593 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477ecb7c-f74b-4f17-87c6-e25d129aaacb-config\") pod \"kube-controller-manager-operator-78b949d7b-qq8rm\" (UID: \"477ecb7c-f74b-4f17-87c6-e25d129aaacb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387618 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-trusted-ca\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387758 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477ecb7c-f74b-4f17-87c6-e25d129aaacb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qq8rm\" (UID: \"477ecb7c-f74b-4f17-87c6-e25d129aaacb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387780 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjngg\" (UniqueName: \"kubernetes.io/projected/3e950818-24fa-4cbe-bdd9-1dbb895b12f1-kube-api-access-jjngg\") pod \"dns-default-x2rjg\" (UID: \"3e950818-24fa-4cbe-bdd9-1dbb895b12f1\") " pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387812 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/03505445-42fb-4015-97d4-6d41167cadd1-default-certificate\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387830 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsdcw\" (UniqueName: \"kubernetes.io/projected/7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb-kube-api-access-fsdcw\") pod \"package-server-manager-789f6589d5-g8rqs\" (UID: \"7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.387849 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-578sl\" (UniqueName: \"kubernetes.io/projected/056a8276-f8ee-45f5-a145-63db980dc0c4-kube-api-access-578sl\") pod \"control-plane-machine-set-operator-78cbb6b69f-8srq6\" (UID: \"056a8276-f8ee-45f5-a145-63db980dc0c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" Oct 04 03:05:32 crc kubenswrapper[4770]: E1004 03:05:32.393251 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:32.893228639 +0000 UTC m=+144.185238351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.409912 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.428183 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.435972 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6"] Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.447755 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.476339 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x95pk"] Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.488678 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:32 crc kubenswrapper[4770]: E1004 03:05:32.489036 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:32.988978603 +0000 UTC m=+144.280988315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489213 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/477ecb7c-f74b-4f17-87c6-e25d129aaacb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qq8rm\" (UID: \"477ecb7c-f74b-4f17-87c6-e25d129aaacb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489296 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g8rqs\" (UID: \"7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489369 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9bb6\" (UniqueName: \"kubernetes.io/projected/40fa4f77-7c68-4688-b6b7-234ba1dd1d33-kube-api-access-r9bb6\") pod \"olm-operator-6b444d44fb-h7khz\" (UID: \"40fa4f77-7c68-4688-b6b7-234ba1dd1d33\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489438 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d25eff-feac-415e-9ba2-e34de3c0d866-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nmsxt\" (UID: \"e9d25eff-feac-415e-9ba2-e34de3c0d866\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489507 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6rq\" (UniqueName: \"kubernetes.io/projected/6425e4f2-75c0-44a2-9416-76df1cb69e94-kube-api-access-9n6rq\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489599 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d25eff-feac-415e-9ba2-e34de3c0d866-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nmsxt\" (UID: \"e9d25eff-feac-415e-9ba2-e34de3c0d866\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489672 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/056a8276-f8ee-45f5-a145-63db980dc0c4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8srq6\" (UID: \"056a8276-f8ee-45f5-a145-63db980dc0c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489739 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29mlx\" (UniqueName: \"kubernetes.io/projected/f54b9ac6-dfe8-4502-95c7-f657e8be4b05-kube-api-access-29mlx\") pod \"migrator-59844c95c7-bzk7x\" (UID: \"f54b9ac6-dfe8-4502-95c7-f657e8be4b05\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489813 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d25eff-feac-415e-9ba2-e34de3c0d866-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nmsxt\" (UID: \"e9d25eff-feac-415e-9ba2-e34de3c0d866\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489885 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f922cbae-d663-4a41-96ed-fe20b4bc3abb-srv-cert\") pod \"catalog-operator-68c6474976-s26vf\" (UID: \"f922cbae-d663-4a41-96ed-fe20b4bc3abb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.489951 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e950818-24fa-4cbe-bdd9-1dbb895b12f1-config-volume\") pod \"dns-default-x2rjg\" (UID: \"3e950818-24fa-4cbe-bdd9-1dbb895b12f1\") " pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490032 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/03505445-42fb-4015-97d4-6d41167cadd1-stats-auth\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490111 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/61ef5869-de69-41b8-a0e7-5da643b24cc3-signing-cabundle\") pod \"service-ca-9c57cc56f-jt2cp\" (UID: \"61ef5869-de69-41b8-a0e7-5da643b24cc3\") " pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490186 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-bound-sa-token\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490288 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-registry-tls\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490355 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/108c220a-bc76-4e17-884d-80f2a5b63357-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490428 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0035edd0-6338-48f1-9534-3b80e135f956-proxy-tls\") pod \"machine-config-controller-84d6567774-rgdpf\" (UID: \"0035edd0-6338-48f1-9534-3b80e135f956\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490491 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e111a15d-a0c1-470b-9723-e4bbcc2265c1-config-volume\") pod \"collect-profiles-29325780-z72ph\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490559 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1c205f2-c2fc-49b5-8f08-7d8173506815-config\") pod \"service-ca-operator-777779d784-ntlrg\" (UID: \"b1c205f2-c2fc-49b5-8f08-7d8173506815\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490624 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2fqvm\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490689 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477ecb7c-f74b-4f17-87c6-e25d129aaacb-config\") pod \"kube-controller-manager-operator-78b949d7b-qq8rm\" (UID: \"477ecb7c-f74b-4f17-87c6-e25d129aaacb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490752 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/61ef5869-de69-41b8-a0e7-5da643b24cc3-signing-key\") pod \"service-ca-9c57cc56f-jt2cp\" (UID: \"61ef5869-de69-41b8-a0e7-5da643b24cc3\") " pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.490918 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-trusted-ca\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.491039 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-mountpoint-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.491114 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477ecb7c-f74b-4f17-87c6-e25d129aaacb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qq8rm\" (UID: \"477ecb7c-f74b-4f17-87c6-e25d129aaacb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.491193 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj2jb\" (UniqueName: \"kubernetes.io/projected/dac89ce4-6624-4d5b-a00f-5c6ffa342535-kube-api-access-wj2jb\") pod \"marketplace-operator-79b997595-2fqvm\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.491289 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d60e800e-4ccc-4db9-ade1-4a8f43775198-cert\") pod \"ingress-canary-twts6\" (UID: \"d60e800e-4ccc-4db9-ade1-4a8f43775198\") " pod="openshift-ingress-canary/ingress-canary-twts6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.491379 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjngg\" (UniqueName: \"kubernetes.io/projected/3e950818-24fa-4cbe-bdd9-1dbb895b12f1-kube-api-access-jjngg\") pod \"dns-default-x2rjg\" (UID: \"3e950818-24fa-4cbe-bdd9-1dbb895b12f1\") " pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.491545 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/40fa4f77-7c68-4688-b6b7-234ba1dd1d33-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h7khz\" (UID: \"40fa4f77-7c68-4688-b6b7-234ba1dd1d33\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492643 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e111a15d-a0c1-470b-9723-e4bbcc2265c1-config-volume\") pod \"collect-profiles-29325780-z72ph\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492664 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/03505445-42fb-4015-97d4-6d41167cadd1-default-certificate\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492714 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsdcw\" (UniqueName: \"kubernetes.io/projected/7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb-kube-api-access-fsdcw\") pod \"package-server-manager-789f6589d5-g8rqs\" (UID: \"7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492747 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-578sl\" (UniqueName: \"kubernetes.io/projected/056a8276-f8ee-45f5-a145-63db980dc0c4-kube-api-access-578sl\") pod \"control-plane-machine-set-operator-78cbb6b69f-8srq6\" (UID: \"056a8276-f8ee-45f5-a145-63db980dc0c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-registry-certificates\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492804 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a5443b52-3d7f-4ff3-9c01-163da49074e2-tmpfs\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492832 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f922cbae-d663-4a41-96ed-fe20b4bc3abb-profile-collector-cert\") pod \"catalog-operator-68c6474976-s26vf\" (UID: \"f922cbae-d663-4a41-96ed-fe20b4bc3abb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492851 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2gh7\" (UniqueName: \"kubernetes.io/projected/f922cbae-d663-4a41-96ed-fe20b4bc3abb-kube-api-access-w2gh7\") pod \"catalog-operator-68c6474976-s26vf\" (UID: \"f922cbae-d663-4a41-96ed-fe20b4bc3abb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492874 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57grp\" (UniqueName: \"kubernetes.io/projected/d60e800e-4ccc-4db9-ade1-4a8f43775198-kube-api-access-57grp\") pod \"ingress-canary-twts6\" (UID: \"d60e800e-4ccc-4db9-ade1-4a8f43775198\") " pod="openshift-ingress-canary/ingress-canary-twts6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492896 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv5lj\" (UniqueName: \"kubernetes.io/projected/b1c205f2-c2fc-49b5-8f08-7d8173506815-kube-api-access-mv5lj\") pod \"service-ca-operator-777779d784-ntlrg\" (UID: \"b1c205f2-c2fc-49b5-8f08-7d8173506815\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492924 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14ca55a-159f-48e5-a290-5462a4f8811c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fxv82\" (UID: \"e14ca55a-159f-48e5-a290-5462a4f8811c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492553 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477ecb7c-f74b-4f17-87c6-e25d129aaacb-config\") pod \"kube-controller-manager-operator-78b949d7b-qq8rm\" (UID: \"477ecb7c-f74b-4f17-87c6-e25d129aaacb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492945 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2g8l\" (UniqueName: \"kubernetes.io/projected/03505445-42fb-4015-97d4-6d41167cadd1-kube-api-access-b2g8l\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.492972 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-registration-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493000 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnrz7\" (UniqueName: \"kubernetes.io/projected/61ef5869-de69-41b8-a0e7-5da643b24cc3-kube-api-access-xnrz7\") pod \"service-ca-9c57cc56f-jt2cp\" (UID: \"61ef5869-de69-41b8-a0e7-5da643b24cc3\") " pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493048 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0e21b999-801f-4c26-9b3f-ac4414511538-certs\") pod \"machine-config-server-v57fn\" (UID: \"0e21b999-801f-4c26-9b3f-ac4414511538\") " pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493071 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-csi-data-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493091 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-trusted-ca\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493127 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5443b52-3d7f-4ff3-9c01-163da49074e2-webhook-cert\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493167 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e14ca55a-159f-48e5-a290-5462a4f8811c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fxv82\" (UID: \"e14ca55a-159f-48e5-a290-5462a4f8811c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493192 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ghlx\" (UniqueName: \"kubernetes.io/projected/e14ca55a-159f-48e5-a290-5462a4f8811c-kube-api-access-2ghlx\") pod \"kube-storage-version-migrator-operator-b67b599dd-fxv82\" (UID: \"e14ca55a-159f-48e5-a290-5462a4f8811c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493223 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03505445-42fb-4015-97d4-6d41167cadd1-metrics-certs\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493240 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5443b52-3d7f-4ff3-9c01-163da49074e2-apiservice-cert\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493295 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/40fa4f77-7c68-4688-b6b7-234ba1dd1d33-srv-cert\") pod \"olm-operator-6b444d44fb-h7khz\" (UID: \"40fa4f77-7c68-4688-b6b7-234ba1dd1d33\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493341 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-plugins-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493376 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/108c220a-bc76-4e17-884d-80f2a5b63357-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493423 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cztnz\" (UniqueName: \"kubernetes.io/projected/e111a15d-a0c1-470b-9723-e4bbcc2265c1-kube-api-access-cztnz\") pod \"collect-profiles-29325780-z72ph\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493453 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-socket-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493475 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03505445-42fb-4015-97d4-6d41167cadd1-service-ca-bundle\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493502 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2fqvm\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493566 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3e950818-24fa-4cbe-bdd9-1dbb895b12f1-metrics-tls\") pod \"dns-default-x2rjg\" (UID: \"3e950818-24fa-4cbe-bdd9-1dbb895b12f1\") " pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493588 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2cnr\" (UniqueName: \"kubernetes.io/projected/0e21b999-801f-4c26-9b3f-ac4414511538-kube-api-access-c2cnr\") pod \"machine-config-server-v57fn\" (UID: \"0e21b999-801f-4c26-9b3f-ac4414511538\") " pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.493624 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494032 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/108c220a-bc76-4e17-884d-80f2a5b63357-ca-trust-extracted\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494055 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-registry-certificates\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494352 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a5443b52-3d7f-4ff3-9c01-163da49074e2-tmpfs\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494577 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03505445-42fb-4015-97d4-6d41167cadd1-service-ca-bundle\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494806 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e111a15d-a0c1-470b-9723-e4bbcc2265c1-secret-volume\") pod \"collect-profiles-29325780-z72ph\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494839 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxsv9\" (UniqueName: \"kubernetes.io/projected/a5443b52-3d7f-4ff3-9c01-163da49074e2-kube-api-access-sxsv9\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494861 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0035edd0-6338-48f1-9534-3b80e135f956-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rgdpf\" (UID: \"0035edd0-6338-48f1-9534-3b80e135f956\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494892 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1c205f2-c2fc-49b5-8f08-7d8173506815-serving-cert\") pod \"service-ca-operator-777779d784-ntlrg\" (UID: \"b1c205f2-c2fc-49b5-8f08-7d8173506815\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494955 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0e21b999-801f-4c26-9b3f-ac4414511538-node-bootstrap-token\") pod \"machine-config-server-v57fn\" (UID: \"0e21b999-801f-4c26-9b3f-ac4414511538\") " pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.494980 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9dmn\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-kube-api-access-j9dmn\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.495456 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d25eff-feac-415e-9ba2-e34de3c0d866-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nmsxt\" (UID: \"e9d25eff-feac-415e-9ba2-e34de3c0d866\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.495976 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14ca55a-159f-48e5-a290-5462a4f8811c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fxv82\" (UID: \"e14ca55a-159f-48e5-a290-5462a4f8811c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: E1004 03:05:32.496247 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:32.996232228 +0000 UTC m=+144.288241940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.496881 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f922cbae-d663-4a41-96ed-fe20b4bc3abb-profile-collector-cert\") pod \"catalog-operator-68c6474976-s26vf\" (UID: \"f922cbae-d663-4a41-96ed-fe20b4bc3abb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.496941 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs84j\" (UniqueName: \"kubernetes.io/projected/0035edd0-6338-48f1-9534-3b80e135f956-kube-api-access-rs84j\") pod \"machine-config-controller-84d6567774-rgdpf\" (UID: \"0035edd0-6338-48f1-9534-3b80e135f956\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.497797 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0035edd0-6338-48f1-9534-3b80e135f956-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rgdpf\" (UID: \"0035edd0-6338-48f1-9534-3b80e135f956\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.497917 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e14ca55a-159f-48e5-a290-5462a4f8811c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fxv82\" (UID: \"e14ca55a-159f-48e5-a290-5462a4f8811c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.498337 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477ecb7c-f74b-4f17-87c6-e25d129aaacb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qq8rm\" (UID: \"477ecb7c-f74b-4f17-87c6-e25d129aaacb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.498510 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5443b52-3d7f-4ff3-9c01-163da49074e2-webhook-cert\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.498768 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5443b52-3d7f-4ff3-9c01-163da49074e2-apiservice-cert\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.500209 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0035edd0-6338-48f1-9534-3b80e135f956-proxy-tls\") pod \"machine-config-controller-84d6567774-rgdpf\" (UID: \"0035edd0-6338-48f1-9534-3b80e135f956\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.506853 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d25eff-feac-415e-9ba2-e34de3c0d866-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nmsxt\" (UID: \"e9d25eff-feac-415e-9ba2-e34de3c0d866\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.506882 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3e950818-24fa-4cbe-bdd9-1dbb895b12f1-metrics-tls\") pod \"dns-default-x2rjg\" (UID: \"3e950818-24fa-4cbe-bdd9-1dbb895b12f1\") " pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.506926 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/03505445-42fb-4015-97d4-6d41167cadd1-default-certificate\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.507533 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-g8rqs\" (UID: \"7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.507620 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/056a8276-f8ee-45f5-a145-63db980dc0c4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8srq6\" (UID: \"056a8276-f8ee-45f5-a145-63db980dc0c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.508485 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03505445-42fb-4015-97d4-6d41167cadd1-metrics-certs\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.509871 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-registry-tls\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.511106 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e111a15d-a0c1-470b-9723-e4bbcc2265c1-secret-volume\") pod \"collect-profiles-29325780-z72ph\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.511447 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f922cbae-d663-4a41-96ed-fe20b4bc3abb-srv-cert\") pod \"catalog-operator-68c6474976-s26vf\" (UID: \"f922cbae-d663-4a41-96ed-fe20b4bc3abb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.512441 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/108c220a-bc76-4e17-884d-80f2a5b63357-installation-pull-secrets\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.514108 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/03505445-42fb-4015-97d4-6d41167cadd1-stats-auth\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.543180 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/477ecb7c-f74b-4f17-87c6-e25d129aaacb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qq8rm\" (UID: \"477ecb7c-f74b-4f17-87c6-e25d129aaacb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.571474 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2g8l\" (UniqueName: \"kubernetes.io/projected/03505445-42fb-4015-97d4-6d41167cadd1-kube-api-access-b2g8l\") pod \"router-default-5444994796-ggg8j\" (UID: \"03505445-42fb-4015-97d4-6d41167cadd1\") " pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.584933 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" event={"ID":"3b6ec72f-1166-4e03-815a-75d256d9ab5a","Type":"ContainerStarted","Data":"3ec90cd73b7a5dfe86bcfbdcd27d0ce9b99241a09686ebcab2c593169a11f79a"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.585040 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" event={"ID":"3b6ec72f-1166-4e03-815a-75d256d9ab5a","Type":"ContainerStarted","Data":"3a6a9ecefb402c1810b76a8e7c8ab5e4eb159d032ec17225de85b0d7fa577874"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.585057 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" event={"ID":"3b6ec72f-1166-4e03-815a-75d256d9ab5a","Type":"ContainerStarted","Data":"1fb6a402a60cfdae785c968d2ed6e54a6ff7a54cf40144544e541a1d7ea33f74"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.591237 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsdcw\" (UniqueName: \"kubernetes.io/projected/7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb-kube-api-access-fsdcw\") pod \"package-server-manager-789f6589d5-g8rqs\" (UID: \"7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.591234 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pv4st" event={"ID":"cb105179-5c3e-4ec2-8204-6898e139c938","Type":"ContainerStarted","Data":"9a385081841c51576a5676d833094f2c71039371d50dd58f824da57ce8245e03"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.594388 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" event={"ID":"941a4e7d-5fdb-4642-8d90-9a9cef86e912","Type":"ContainerStarted","Data":"c0bf10788721118367adcd8840a074718ef88ea8267e92802d4b9cbc736f7070"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.596429 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" event={"ID":"0e1a8be1-36ca-4579-9678-13ba76cec6b1","Type":"ContainerStarted","Data":"75d7ac4ed07d80aa5ddec4a2d6531994033874ad6ebc01216236490f08cbc9cf"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.596464 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" event={"ID":"0e1a8be1-36ca-4579-9678-13ba76cec6b1","Type":"ContainerStarted","Data":"aa89aeb7d560709fa9d9900a1c3aff3fee24cd9708f8987dfffccdc44a106157"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.599322 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" event={"ID":"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4","Type":"ContainerStarted","Data":"1543ae92a7ecd0dc9db2ff70ad9d3e2566017c1b91a9cd9cb0d8ea9498bd08c2"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.599843 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:32 crc kubenswrapper[4770]: E1004 03:05:32.600069 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.099999869 +0000 UTC m=+144.392009581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600209 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2cnr\" (UniqueName: \"kubernetes.io/projected/0e21b999-801f-4c26-9b3f-ac4414511538-kube-api-access-c2cnr\") pod \"machine-config-server-v57fn\" (UID: \"0e21b999-801f-4c26-9b3f-ac4414511538\") " pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600253 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600296 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1c205f2-c2fc-49b5-8f08-7d8173506815-serving-cert\") pod \"service-ca-operator-777779d784-ntlrg\" (UID: \"b1c205f2-c2fc-49b5-8f08-7d8173506815\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600320 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0e21b999-801f-4c26-9b3f-ac4414511538-node-bootstrap-token\") pod \"machine-config-server-v57fn\" (UID: \"0e21b999-801f-4c26-9b3f-ac4414511538\") " pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600366 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9bb6\" (UniqueName: \"kubernetes.io/projected/40fa4f77-7c68-4688-b6b7-234ba1dd1d33-kube-api-access-r9bb6\") pod \"olm-operator-6b444d44fb-h7khz\" (UID: \"40fa4f77-7c68-4688-b6b7-234ba1dd1d33\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600399 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n6rq\" (UniqueName: \"kubernetes.io/projected/6425e4f2-75c0-44a2-9416-76df1cb69e94-kube-api-access-9n6rq\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600489 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/61ef5869-de69-41b8-a0e7-5da643b24cc3-signing-cabundle\") pod \"service-ca-9c57cc56f-jt2cp\" (UID: \"61ef5869-de69-41b8-a0e7-5da643b24cc3\") " pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600551 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1c205f2-c2fc-49b5-8f08-7d8173506815-config\") pod \"service-ca-operator-777779d784-ntlrg\" (UID: \"b1c205f2-c2fc-49b5-8f08-7d8173506815\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600577 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2fqvm\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600603 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/61ef5869-de69-41b8-a0e7-5da643b24cc3-signing-key\") pod \"service-ca-9c57cc56f-jt2cp\" (UID: \"61ef5869-de69-41b8-a0e7-5da643b24cc3\") " pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600639 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-mountpoint-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600665 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj2jb\" (UniqueName: \"kubernetes.io/projected/dac89ce4-6624-4d5b-a00f-5c6ffa342535-kube-api-access-wj2jb\") pod \"marketplace-operator-79b997595-2fqvm\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600689 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d60e800e-4ccc-4db9-ade1-4a8f43775198-cert\") pod \"ingress-canary-twts6\" (UID: \"d60e800e-4ccc-4db9-ade1-4a8f43775198\") " pod="openshift-ingress-canary/ingress-canary-twts6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600720 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/40fa4f77-7c68-4688-b6b7-234ba1dd1d33-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h7khz\" (UID: \"40fa4f77-7c68-4688-b6b7-234ba1dd1d33\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57grp\" (UniqueName: \"kubernetes.io/projected/d60e800e-4ccc-4db9-ade1-4a8f43775198-kube-api-access-57grp\") pod \"ingress-canary-twts6\" (UID: \"d60e800e-4ccc-4db9-ade1-4a8f43775198\") " pod="openshift-ingress-canary/ingress-canary-twts6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600794 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv5lj\" (UniqueName: \"kubernetes.io/projected/b1c205f2-c2fc-49b5-8f08-7d8173506815-kube-api-access-mv5lj\") pod \"service-ca-operator-777779d784-ntlrg\" (UID: \"b1c205f2-c2fc-49b5-8f08-7d8173506815\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600821 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-registration-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600846 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnrz7\" (UniqueName: \"kubernetes.io/projected/61ef5869-de69-41b8-a0e7-5da643b24cc3-kube-api-access-xnrz7\") pod \"service-ca-9c57cc56f-jt2cp\" (UID: \"61ef5869-de69-41b8-a0e7-5da643b24cc3\") " pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600870 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0e21b999-801f-4c26-9b3f-ac4414511538-certs\") pod \"machine-config-server-v57fn\" (UID: \"0e21b999-801f-4c26-9b3f-ac4414511538\") " pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600894 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-csi-data-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600946 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/40fa4f77-7c68-4688-b6b7-234ba1dd1d33-srv-cert\") pod \"olm-operator-6b444d44fb-h7khz\" (UID: \"40fa4f77-7c68-4688-b6b7-234ba1dd1d33\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600968 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-plugins-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.600997 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-socket-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.601055 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2fqvm\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.602206 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-578sl\" (UniqueName: \"kubernetes.io/projected/056a8276-f8ee-45f5-a145-63db980dc0c4-kube-api-access-578sl\") pod \"control-plane-machine-set-operator-78cbb6b69f-8srq6\" (UID: \"056a8276-f8ee-45f5-a145-63db980dc0c4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" Oct 04 03:05:32 crc kubenswrapper[4770]: E1004 03:05:32.602252 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.102243226 +0000 UTC m=+144.394252938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.604667 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2fqvm\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.604905 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-csi-data-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.605662 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1c205f2-c2fc-49b5-8f08-7d8173506815-config\") pod \"service-ca-operator-777779d784-ntlrg\" (UID: \"b1c205f2-c2fc-49b5-8f08-7d8173506815\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.605792 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-socket-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.605812 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-registration-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.606319 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" event={"ID":"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7","Type":"ContainerStarted","Data":"14eb028a5a4408f7009f002bc47892a2b3a4f79cff8fa1dc0fdf791dec1aa47f"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.606428 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-plugins-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.608418 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2fqvm\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.608554 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6425e4f2-75c0-44a2-9416-76df1cb69e94-mountpoint-dir\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.610108 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0e21b999-801f-4c26-9b3f-ac4414511538-certs\") pod \"machine-config-server-v57fn\" (UID: \"0e21b999-801f-4c26-9b3f-ac4414511538\") " pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.610124 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0e21b999-801f-4c26-9b3f-ac4414511538-node-bootstrap-token\") pod \"machine-config-server-v57fn\" (UID: \"0e21b999-801f-4c26-9b3f-ac4414511538\") " pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.610372 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1c205f2-c2fc-49b5-8f08-7d8173506815-serving-cert\") pod \"service-ca-operator-777779d784-ntlrg\" (UID: \"b1c205f2-c2fc-49b5-8f08-7d8173506815\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.610824 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" event={"ID":"c7a98e59-558d-4731-a626-3dc946844eaa","Type":"ContainerStarted","Data":"6e43fc069a72441bb50a5e3b6a2a77fa8a575e04484530b8c8120418b27f3fff"} Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.611135 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/40fa4f77-7c68-4688-b6b7-234ba1dd1d33-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h7khz\" (UID: \"40fa4f77-7c68-4688-b6b7-234ba1dd1d33\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.613479 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d60e800e-4ccc-4db9-ade1-4a8f43775198-cert\") pod \"ingress-canary-twts6\" (UID: \"d60e800e-4ccc-4db9-ade1-4a8f43775198\") " pod="openshift-ingress-canary/ingress-canary-twts6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.617524 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/40fa4f77-7c68-4688-b6b7-234ba1dd1d33-srv-cert\") pod \"olm-operator-6b444d44fb-h7khz\" (UID: \"40fa4f77-7c68-4688-b6b7-234ba1dd1d33\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.619240 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d25eff-feac-415e-9ba2-e34de3c0d866-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-nmsxt\" (UID: \"e9d25eff-feac-415e-9ba2-e34de3c0d866\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.647747 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cztnz\" (UniqueName: \"kubernetes.io/projected/e111a15d-a0c1-470b-9723-e4bbcc2265c1-kube-api-access-cztnz\") pod \"collect-profiles-29325780-z72ph\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.663607 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ghlx\" (UniqueName: \"kubernetes.io/projected/e14ca55a-159f-48e5-a290-5462a4f8811c-kube-api-access-2ghlx\") pod \"kube-storage-version-migrator-operator-b67b599dd-fxv82\" (UID: \"e14ca55a-159f-48e5-a290-5462a4f8811c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.677050 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjngg\" (UniqueName: \"kubernetes.io/projected/3e950818-24fa-4cbe-bdd9-1dbb895b12f1-kube-api-access-jjngg\") pod \"dns-default-x2rjg\" (UID: \"3e950818-24fa-4cbe-bdd9-1dbb895b12f1\") " pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.686933 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/61ef5869-de69-41b8-a0e7-5da643b24cc3-signing-cabundle\") pod \"service-ca-9c57cc56f-jt2cp\" (UID: \"61ef5869-de69-41b8-a0e7-5da643b24cc3\") " pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.687593 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e950818-24fa-4cbe-bdd9-1dbb895b12f1-config-volume\") pod \"dns-default-x2rjg\" (UID: \"3e950818-24fa-4cbe-bdd9-1dbb895b12f1\") " pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.687966 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/61ef5869-de69-41b8-a0e7-5da643b24cc3-signing-key\") pod \"service-ca-9c57cc56f-jt2cp\" (UID: \"61ef5869-de69-41b8-a0e7-5da643b24cc3\") " pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.701966 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:32 crc kubenswrapper[4770]: E1004 03:05:32.702530 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.202511467 +0000 UTC m=+144.494521179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.716717 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2gh7\" (UniqueName: \"kubernetes.io/projected/f922cbae-d663-4a41-96ed-fe20b4bc3abb-kube-api-access-w2gh7\") pod \"catalog-operator-68c6474976-s26vf\" (UID: \"f922cbae-d663-4a41-96ed-fe20b4bc3abb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.726247 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29mlx\" (UniqueName: \"kubernetes.io/projected/f54b9ac6-dfe8-4502-95c7-f657e8be4b05-kube-api-access-29mlx\") pod \"migrator-59844c95c7-bzk7x\" (UID: \"f54b9ac6-dfe8-4502-95c7-f657e8be4b05\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.745643 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-bound-sa-token\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.759575 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9dmn\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-kube-api-access-j9dmn\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.780609 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxsv9\" (UniqueName: \"kubernetes.io/projected/a5443b52-3d7f-4ff3-9c01-163da49074e2-kube-api-access-sxsv9\") pod \"packageserver-d55dfcdfc-ssbgd\" (UID: \"a5443b52-3d7f-4ff3-9c01-163da49074e2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.789229 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.799424 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.801203 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mwtcd"] Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.803104 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: E1004 03:05:32.803498 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.303485675 +0000 UTC m=+144.595495387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.807523 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.809949 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs84j\" (UniqueName: \"kubernetes.io/projected/0035edd0-6338-48f1-9534-3b80e135f956-kube-api-access-rs84j\") pod \"machine-config-controller-84d6567774-rgdpf\" (UID: \"0035edd0-6338-48f1-9534-3b80e135f956\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.814308 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.824183 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.830516 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.838898 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.845764 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n6rq\" (UniqueName: \"kubernetes.io/projected/6425e4f2-75c0-44a2-9416-76df1cb69e94-kube-api-access-9n6rq\") pod \"csi-hostpathplugin-vxv67\" (UID: \"6425e4f2-75c0-44a2-9416-76df1cb69e94\") " pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.846138 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.853182 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.869156 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.870613 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-kpt6s"] Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.875740 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.881347 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2cnr\" (UniqueName: \"kubernetes.io/projected/0e21b999-801f-4c26-9b3f-ac4414511538-kube-api-access-c2cnr\") pod \"machine-config-server-v57fn\" (UID: \"0e21b999-801f-4c26-9b3f-ac4414511538\") " pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.883690 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr"] Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.885223 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9"] Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.885430 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.908380 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj2jb\" (UniqueName: \"kubernetes.io/projected/dac89ce4-6624-4d5b-a00f-5c6ffa342535-kube-api-access-wj2jb\") pod \"marketplace-operator-79b997595-2fqvm\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.908864 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:32 crc kubenswrapper[4770]: E1004 03:05:32.912493 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.412459758 +0000 UTC m=+144.704469470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.912924 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnrz7\" (UniqueName: \"kubernetes.io/projected/61ef5869-de69-41b8-a0e7-5da643b24cc3-kube-api-access-xnrz7\") pod \"service-ca-9c57cc56f-jt2cp\" (UID: \"61ef5869-de69-41b8-a0e7-5da643b24cc3\") " pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.914589 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xkkrp"] Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.928171 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57grp\" (UniqueName: \"kubernetes.io/projected/d60e800e-4ccc-4db9-ade1-4a8f43775198-kube-api-access-57grp\") pod \"ingress-canary-twts6\" (UID: \"d60e800e-4ccc-4db9-ade1-4a8f43775198\") " pod="openshift-ingress-canary/ingress-canary-twts6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.931421 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:32 crc kubenswrapper[4770]: E1004 03:05:32.931862 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.431848933 +0000 UTC m=+144.723858635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.935163 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vxv67" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.943052 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-twts6" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.948291 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-v57fn" Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.966212 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9bb6\" (UniqueName: \"kubernetes.io/projected/40fa4f77-7c68-4688-b6b7-234ba1dd1d33-kube-api-access-r9bb6\") pod \"olm-operator-6b444d44fb-h7khz\" (UID: \"40fa4f77-7c68-4688-b6b7-234ba1dd1d33\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:32 crc kubenswrapper[4770]: W1004 03:05:32.975184 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61545ae7_a4d3_49dc_95a4_6393143a3718.slice/crio-3a5310b8155a6275970f3733f07cb30ea0daf4d31a08f07d5cdb441ce0b39aa1 WatchSource:0}: Error finding container 3a5310b8155a6275970f3733f07cb30ea0daf4d31a08f07d5cdb441ce0b39aa1: Status 404 returned error can't find the container with id 3a5310b8155a6275970f3733f07cb30ea0daf4d31a08f07d5cdb441ce0b39aa1 Oct 04 03:05:32 crc kubenswrapper[4770]: I1004 03:05:32.978817 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv5lj\" (UniqueName: \"kubernetes.io/projected/b1c205f2-c2fc-49b5-8f08-7d8173506815-kube-api-access-mv5lj\") pod \"service-ca-operator-777779d784-ntlrg\" (UID: \"b1c205f2-c2fc-49b5-8f08-7d8173506815\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.035909 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.036043 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.536001983 +0000 UTC m=+144.828011695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.036984 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.037329 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.537321356 +0000 UTC m=+144.829331068 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.038132 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ppsgj"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.058372 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.065578 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r8rk"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.080784 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.115481 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.118738 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.126975 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xwhz4"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.129106 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lwhtk"] Oct 04 03:05:33 crc kubenswrapper[4770]: W1004 03:05:33.137286 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03505445_42fb_4015_97d4_6d41167cadd1.slice/crio-120490ee10cef9b996c7e8a0ec442be13e5f4b314c292fac148e82e55255a4b2 WatchSource:0}: Error finding container 120490ee10cef9b996c7e8a0ec442be13e5f4b314c292fac148e82e55255a4b2: Status 404 returned error can't find the container with id 120490ee10cef9b996c7e8a0ec442be13e5f4b314c292fac148e82e55255a4b2 Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.137800 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.138229 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.638209163 +0000 UTC m=+144.930218875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: W1004 03:05:33.171497 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda68c8d9f_b526_4583_901a_23595a0704a1.slice/crio-04f7c8fd47f02062203aa93f39e33e982ceb48dae863ba8562e853341b650108 WatchSource:0}: Error finding container 04f7c8fd47f02062203aa93f39e33e982ceb48dae863ba8562e853341b650108: Status 404 returned error can't find the container with id 04f7c8fd47f02062203aa93f39e33e982ceb48dae863ba8562e853341b650108 Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.187162 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.194761 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.199714 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.209967 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.215658 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.239199 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.240173 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.740161606 +0000 UTC m=+145.032171318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.340125 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.340311 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.840282352 +0000 UTC m=+145.132292064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.342952 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.343527 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.843512366 +0000 UTC m=+145.135522078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.393077 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.446309 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.449153 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:33.949131042 +0000 UTC m=+145.241140764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.489773 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf"] Oct 04 03:05:33 crc kubenswrapper[4770]: W1004 03:05:33.517034 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf54b9ac6_dfe8_4502_95c7_f657e8be4b05.slice/crio-261bb67a76a6449928d4d2c1b5c9acd25ca9558bf7f55a1c0c60a126503c51d5 WatchSource:0}: Error finding container 261bb67a76a6449928d4d2c1b5c9acd25ca9558bf7f55a1c0c60a126503c51d5: Status 404 returned error can't find the container with id 261bb67a76a6449928d4d2c1b5c9acd25ca9558bf7f55a1c0c60a126503c51d5 Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.550137 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.550649 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.050627113 +0000 UTC m=+145.342636825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.626231 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-v57fn" event={"ID":"0e21b999-801f-4c26-9b3f-ac4414511538","Type":"ContainerStarted","Data":"6dabb6ef4c8c5f6bd2e0970df5a3555bfdecf8c6f1d7e2047715e6f4b58187aa"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.629854 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" event={"ID":"e5c5896d-a1f2-433c-858d-cea502c067d8","Type":"ContainerStarted","Data":"43fbb3e54d60896516ba64c28da1842364fcea25ede3f940e2231d535f6f1bce"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.635217 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" event={"ID":"45c9685e-2941-4778-97d2-f745955fa577","Type":"ContainerStarted","Data":"15b7814bdacbb602144f70721f0934dbb8bb99dc3511897a7625906f6e00b5bf"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.651689 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.651974 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.151954242 +0000 UTC m=+145.443963954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.653763 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" event={"ID":"b1fc24e8-c0cc-499d-8498-bf591b24ee64","Type":"ContainerStarted","Data":"4440016f7c33f3b29d990b7a3c719281ca2b0b45019d16c12dceba1626bef4ba"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.658852 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" event={"ID":"b153ca4e-5d53-442d-8a73-72ff80ebbe1c","Type":"ContainerStarted","Data":"32a8642b971183615bed2a1f449ca999ba94b5d49337c29e01f598f2f63df9ac"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.692957 4770 generic.go:334] "Generic (PLEG): container finished" podID="d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7" containerID="a99eae2f76ead28d70b81b953b283664a8d16046c8c0628c5c67e9b2007d2ab7" exitCode=0 Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.694582 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" event={"ID":"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7","Type":"ContainerDied","Data":"a99eae2f76ead28d70b81b953b283664a8d16046c8c0628c5c67e9b2007d2ab7"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.698123 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.723209 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" event={"ID":"a1483904-9014-4f9d-944b-3612fbd3b3d9","Type":"ContainerStarted","Data":"5398ab6bbf37f7249c595d6be58131a5b83f22fbc7184e2b560e2e47db153171"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.757066 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.757911 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.257892937 +0000 UTC m=+145.549902649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.759489 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" event={"ID":"c7a98e59-558d-4731-a626-3dc946844eaa","Type":"ContainerStarted","Data":"05b306f12cfa6f1b2e619b7d292de90e3280592929cde6b4d89e190d9c2090d0"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.770245 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.813854 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pv4st" event={"ID":"cb105179-5c3e-4ec2-8204-6898e139c938","Type":"ContainerStarted","Data":"4ed9688fc476d9f115d9f71adfbc08c7defdf4cca6e99e8ecd14599993835a95"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.816862 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.819060 4770 generic.go:334] "Generic (PLEG): container finished" podID="941a4e7d-5fdb-4642-8d90-9a9cef86e912" containerID="6af679313d7ec14c4f9067bb030d0704bb930fa1bae529916f774ae9ad1e970e" exitCode=0 Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.819364 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" event={"ID":"941a4e7d-5fdb-4642-8d90-9a9cef86e912","Type":"ContainerDied","Data":"6af679313d7ec14c4f9067bb030d0704bb930fa1bae529916f774ae9ad1e970e"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.829412 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" event={"ID":"253c36e7-4eba-492f-8057-fd3865fcd223","Type":"ContainerStarted","Data":"a0abee0085e5ab79c8be04edeb40e6e5915b55eb6e32f26ed28837c42018d300"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.831417 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" event={"ID":"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4","Type":"ContainerStarted","Data":"73b895e5f3390592e5a1eba21de72b3d9cecbf3cecf19ca3d185371a2296417d"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.832626 4770 patch_prober.go:28] interesting pod/console-operator-58897d9998-pv4st container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.832671 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-pv4st" podUID="cb105179-5c3e-4ec2-8204-6898e139c938" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.836612 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" event={"ID":"9070d4f9-0fc3-4542-b010-b83193b54a0d","Type":"ContainerStarted","Data":"138a2746b33ff636b84360fda7d38319afabc085081bc2b8188f60b7b95a30d2"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.841253 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ggg8j" event={"ID":"03505445-42fb-4015-97d4-6d41167cadd1","Type":"ContainerStarted","Data":"120490ee10cef9b996c7e8a0ec442be13e5f4b314c292fac148e82e55255a4b2"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.845318 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" event={"ID":"c4c445d6-f62c-4d63-8d18-35441eebeb49","Type":"ContainerStarted","Data":"b2ffb19386f5c458934aa45f349121fd237e7cdfe843c2198c7e1bd5c8c576b0"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.862502 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.864528 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.364505149 +0000 UTC m=+145.656514861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.874220 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" event={"ID":"0e1a8be1-36ca-4579-9678-13ba76cec6b1","Type":"ContainerStarted","Data":"dbe040ed43ce94f6dc5d50120997e9e0e477c10e443e80fd89f8bb1ebff22090"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.894693 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" event={"ID":"a68c8d9f-b526-4583-901a-23595a0704a1","Type":"ContainerStarted","Data":"04f7c8fd47f02062203aa93f39e33e982ceb48dae863ba8562e853341b650108"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.901367 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.915912 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vxv67"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.919915 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.930050 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd"] Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.931168 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" event={"ID":"42188dc5-d97c-423a-8562-4e7ece931916","Type":"ContainerStarted","Data":"960e63445acae2f25d9fa2bcf0e7912e6132f90a00402673e462bd305fbb6f9e"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.934473 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lwhtk" event={"ID":"4bf7a322-09ed-493a-bd28-2639f15402f5","Type":"ContainerStarted","Data":"058841c1810c326b0831f0fd9d756eb6cf5794e6f3deea98375ca2bb86df093e"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.939887 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" event={"ID":"0a7275fe-266a-4920-a30d-28d3f0ee4a03","Type":"ContainerStarted","Data":"a46d70981b3142e90eae9d51056fd3d28256116716e87a2b822dd83fc2914aa1"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.939934 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" event={"ID":"0a7275fe-266a-4920-a30d-28d3f0ee4a03","Type":"ContainerStarted","Data":"36c6bbf513c7851b11664b4180694b3b4acbfe75a83a1d9b42fdc9b61d11fcc7"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.945649 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xwhz4" event={"ID":"2b45833c-2e11-428e-8131-1ea4676a794f","Type":"ContainerStarted","Data":"9f3292fbfd3849da26529499e7e7fe3384b2e23d9bb336a97a84721af063c3f7"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.951706 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x" event={"ID":"f54b9ac6-dfe8-4502-95c7-f657e8be4b05","Type":"ContainerStarted","Data":"261bb67a76a6449928d4d2c1b5c9acd25ca9558bf7f55a1c0c60a126503c51d5"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.959910 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" event={"ID":"ec292e26-13cf-49b1-83a5-1c325bdd6d99","Type":"ContainerStarted","Data":"a8888c882a3cb5ad31e0a74e298ae096d8d5b35639cf6ccaad23972783c6dd3a"} Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.964436 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:33 crc kubenswrapper[4770]: E1004 03:05:33.965461 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.465442997 +0000 UTC m=+145.757452709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:33 crc kubenswrapper[4770]: I1004 03:05:33.984799 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" event={"ID":"61545ae7-a4d3-49dc-95a4-6393143a3718","Type":"ContainerStarted","Data":"3a5310b8155a6275970f3733f07cb30ea0daf4d31a08f07d5cdb441ce0b39aa1"} Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.078771 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.080320 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.58030002 +0000 UTC m=+145.872309732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.104952 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x2rjg"] Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.113252 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-twts6"] Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.117275 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs"] Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.141420 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.182124 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.182445 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.682431548 +0000 UTC m=+145.974441260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.183978 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2fqvm"] Oct 04 03:05:34 crc kubenswrapper[4770]: W1004 03:05:34.193547 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9d25eff_feac_415e_9ba2_e34de3c0d866.slice/crio-efacc3a349070f974ba4dac8d45d01d216b7ade61e8c1df4e23b5f7d0ab49d39 WatchSource:0}: Error finding container efacc3a349070f974ba4dac8d45d01d216b7ade61e8c1df4e23b5f7d0ab49d39: Status 404 returned error can't find the container with id efacc3a349070f974ba4dac8d45d01d216b7ade61e8c1df4e23b5f7d0ab49d39 Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.215169 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6"] Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.216954 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph"] Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.222589 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jt2cp"] Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.262170 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg"] Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.283475 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.283714 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.783685843 +0000 UTC m=+146.075695555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.284230 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.284606 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.784592257 +0000 UTC m=+146.076601959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.295369 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nwxt6" podStartSLOduration=124.295347291 podStartE2EDuration="2m4.295347291s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:34.291964825 +0000 UTC m=+145.583974537" watchObservedRunningTime="2025-10-04 03:05:34.295347291 +0000 UTC m=+145.587357003" Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.344728 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz"] Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.375091 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" podStartSLOduration=124.375066366 podStartE2EDuration="2m4.375066366s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:34.374471372 +0000 UTC m=+145.666481084" watchObservedRunningTime="2025-10-04 03:05:34.375066366 +0000 UTC m=+145.667076098" Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.386800 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.387231 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.887213797 +0000 UTC m=+146.179223509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: W1004 03:05:34.405544 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61ef5869_de69_41b8_a0e7_5da643b24cc3.slice/crio-af6f5f28958e75163170ad023be55814bc9372c226c231c98b76d1023ad79620 WatchSource:0}: Error finding container af6f5f28958e75163170ad023be55814bc9372c226c231c98b76d1023ad79620: Status 404 returned error can't find the container with id af6f5f28958e75163170ad023be55814bc9372c226c231c98b76d1023ad79620 Oct 04 03:05:34 crc kubenswrapper[4770]: W1004 03:05:34.416182 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40fa4f77_7c68_4688_b6b7_234ba1dd1d33.slice/crio-7f0238b866d22cc94a0e18d64ceae99011aa3197d64859fed0c44a518fbbbeb2 WatchSource:0}: Error finding container 7f0238b866d22cc94a0e18d64ceae99011aa3197d64859fed0c44a518fbbbeb2: Status 404 returned error can't find the container with id 7f0238b866d22cc94a0e18d64ceae99011aa3197d64859fed0c44a518fbbbeb2 Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.463933 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-pv4st" podStartSLOduration=124.463912226 podStartE2EDuration="2m4.463912226s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:34.463784403 +0000 UTC m=+145.755794125" watchObservedRunningTime="2025-10-04 03:05:34.463912226 +0000 UTC m=+145.755921938" Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.494554 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.495171 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:34.995158773 +0000 UTC m=+146.287168485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.516968 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbjnf" podStartSLOduration=123.51694784 podStartE2EDuration="2m3.51694784s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:34.510621748 +0000 UTC m=+145.802631460" watchObservedRunningTime="2025-10-04 03:05:34.51694784 +0000 UTC m=+145.808957552" Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.595855 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.596507 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.09647434 +0000 UTC m=+146.388484052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.596739 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.597199 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.097184579 +0000 UTC m=+146.389194291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.702966 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.703130 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.203097354 +0000 UTC m=+146.495107066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.703689 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.704117 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.204091568 +0000 UTC m=+146.496101320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.807765 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.808437 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.308413453 +0000 UTC m=+146.600423165 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:34 crc kubenswrapper[4770]: I1004 03:05:34.909389 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:34 crc kubenswrapper[4770]: E1004 03:05:34.909837 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.409817862 +0000 UTC m=+146.701827574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.010871 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.011346 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.511318684 +0000 UTC m=+146.803328406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.023874 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" event={"ID":"b1c205f2-c2fc-49b5-8f08-7d8173506815","Type":"ContainerStarted","Data":"533b9c757493245e3821c6288971382248431ad0288ae9292bf6c0a1b5293a3f"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.034633 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" event={"ID":"7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb","Type":"ContainerStarted","Data":"4e97824460a62f90b62d32ed6827a54c8c40b103e63a85f33fde0e7691b0adea"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.041985 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" event={"ID":"056a8276-f8ee-45f5-a145-63db980dc0c4","Type":"ContainerStarted","Data":"58c82585c8ea1f380d8b0f5afff808b8fccbbf470341501ed0006d0e435cc4dd"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.047980 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" event={"ID":"e14ca55a-159f-48e5-a290-5462a4f8811c","Type":"ContainerStarted","Data":"c314349ede3d7038626a0c960750e5932e1251cce6979a2f4f74cfe725a5c12e"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.060274 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ggg8j" event={"ID":"03505445-42fb-4015-97d4-6d41167cadd1","Type":"ContainerStarted","Data":"3b9cca26163c5c056984970ea5b86cd978046274b8e165ec6f0d738968c56049"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.078893 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" event={"ID":"2ee857f0-4ff7-4c3a-b0cb-5f6f4345a4d4","Type":"ContainerStarted","Data":"9068a501a1c64f4d8e6729a2df7c5d55904c4120cc17b060186f799d4160fc79"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.117883 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" event={"ID":"f922cbae-d663-4a41-96ed-fe20b4bc3abb","Type":"ContainerStarted","Data":"dcf91dba4492083e02fbbbe30bbe572aeb5f012182d2f0bc66458ad74a6e7ada"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.117944 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" event={"ID":"f922cbae-d663-4a41-96ed-fe20b4bc3abb","Type":"ContainerStarted","Data":"618ee18fec997d6fc37da48dae3a82f93ef99fa5a2e73ee2647097d19bcce64a"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.118059 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.118780 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.120464 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ggg8j" podStartSLOduration=124.120439911 podStartE2EDuration="2m4.120439911s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.118411159 +0000 UTC m=+146.410420871" watchObservedRunningTime="2025-10-04 03:05:35.120439911 +0000 UTC m=+146.412449643" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.122429 4770 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-s26vf container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.122540 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" podUID="f922cbae-d663-4a41-96ed-fe20b4bc3abb" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.123448 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.623408457 +0000 UTC m=+146.915418169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.147321 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vl" podStartSLOduration=125.147292746 podStartE2EDuration="2m5.147292746s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.143408987 +0000 UTC m=+146.435418699" watchObservedRunningTime="2025-10-04 03:05:35.147292746 +0000 UTC m=+146.439302458" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.155724 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" event={"ID":"253c36e7-4eba-492f-8057-fd3865fcd223","Type":"ContainerStarted","Data":"642e57b9fc195fdc94d38e5a3ba1ba6018f63a7d1b09a1063724143cdfc1b4e8"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.162826 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" event={"ID":"9070d4f9-0fc3-4542-b010-b83193b54a0d","Type":"ContainerStarted","Data":"93eb99815d457ed582dfb0ea55917df892f74ee2a6656df6971633963d608519"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.162868 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" event={"ID":"9070d4f9-0fc3-4542-b010-b83193b54a0d","Type":"ContainerStarted","Data":"dae4899e1068c97000f2aa595ec813eced44c6a01620e7c50a181980f5d13322"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.177556 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" event={"ID":"45c9685e-2941-4778-97d2-f745955fa577","Type":"ContainerStarted","Data":"dd0d47a0ef811a3f1a95564d6f288e0a3dff34b5e663f02dadeeb9aa0a38d763"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.180159 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.189371 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" event={"ID":"e9d25eff-feac-415e-9ba2-e34de3c0d866","Type":"ContainerStarted","Data":"efacc3a349070f974ba4dac8d45d01d216b7ade61e8c1df4e23b5f7d0ab49d39"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.192794 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" podStartSLOduration=124.192765387 podStartE2EDuration="2m4.192765387s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.180490764 +0000 UTC m=+146.472500466" watchObservedRunningTime="2025-10-04 03:05:35.192765387 +0000 UTC m=+146.484775099" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.201534 4770 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6r8rk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.201652 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" podUID="45c9685e-2941-4778-97d2-f745955fa577" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.220885 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.232295 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" podStartSLOduration=124.232273047 podStartE2EDuration="2m4.232273047s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.219948092 +0000 UTC m=+146.511957814" watchObservedRunningTime="2025-10-04 03:05:35.232273047 +0000 UTC m=+146.524282759" Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.234973 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.734944764 +0000 UTC m=+147.026954476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.249484 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" event={"ID":"0a7275fe-266a-4920-a30d-28d3f0ee4a03","Type":"ContainerStarted","Data":"a7dcc343ba0d300bfabbac7dea6d1997f79ca8d570971ee3a40f7ee93f5883a1"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.260501 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-h7lrz" podStartSLOduration=124.260469467 podStartE2EDuration="2m4.260469467s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.252307098 +0000 UTC m=+146.544316810" watchObservedRunningTime="2025-10-04 03:05:35.260469467 +0000 UTC m=+146.552479179" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.330367 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzgxr" podStartSLOduration=124.330341191 podStartE2EDuration="2m4.330341191s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.281566985 +0000 UTC m=+146.573576697" watchObservedRunningTime="2025-10-04 03:05:35.330341191 +0000 UTC m=+146.622350903" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.331703 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.332653 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.832629509 +0000 UTC m=+147.124639221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.355938 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-x95pk" podStartSLOduration=125.355913414 podStartE2EDuration="2m5.355913414s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.332122626 +0000 UTC m=+146.624132328" watchObservedRunningTime="2025-10-04 03:05:35.355913414 +0000 UTC m=+146.647923126" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.359673 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" event={"ID":"0035edd0-6338-48f1-9534-3b80e135f956","Type":"ContainerStarted","Data":"b998024ab67764f3bc98a15500ce365acc44c414481893335f6e2e02feedcfc1"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.359733 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" event={"ID":"0035edd0-6338-48f1-9534-3b80e135f956","Type":"ContainerStarted","Data":"6877d274991ade32cab7c7d1a1b46ad9eea4cf67fed20e9c1826b8eaf3188461"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.361965 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vxv67" event={"ID":"6425e4f2-75c0-44a2-9416-76df1cb69e94","Type":"ContainerStarted","Data":"901ce28c2c3b162ab518ddefac13d31e2ffbcecf4b78839043c79ac6c92d4a74"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.373477 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" event={"ID":"b1fc24e8-c0cc-499d-8498-bf591b24ee64","Type":"ContainerStarted","Data":"d9527e84cb697482f5ed3b74905dd2c5b0dea4ad37742b099219236566309093"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.392916 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x2rjg" event={"ID":"3e950818-24fa-4cbe-bdd9-1dbb895b12f1","Type":"ContainerStarted","Data":"b1cc821ca8c3a2b4d9ad96ca41d8ae4ebabae9ba9a3d76b0dda2f248575ac756"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.408729 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mwtcd" podStartSLOduration=125.408703932 podStartE2EDuration="2m5.408703932s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.408429424 +0000 UTC m=+146.700439136" watchObservedRunningTime="2025-10-04 03:05:35.408703932 +0000 UTC m=+146.700713644" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.433091 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.435182 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:35.935144746 +0000 UTC m=+147.227154458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.444436 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xwhz4" event={"ID":"2b45833c-2e11-428e-8131-1ea4676a794f","Type":"ContainerStarted","Data":"5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.453041 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" event={"ID":"42188dc5-d97c-423a-8562-4e7ece931916","Type":"ContainerStarted","Data":"e8b3db1052282801bd5c44760cb9c8210015498a95d946ce6d8082ae9bd4f11b"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.461563 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x" event={"ID":"f54b9ac6-dfe8-4502-95c7-f657e8be4b05","Type":"ContainerStarted","Data":"3bc2f9427d25e8f4500ffd774994a7fcf16957aa00e50d5fc3cdb2000ca06470"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.476207 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lwhtk" event={"ID":"4bf7a322-09ed-493a-bd28-2639f15402f5","Type":"ContainerStarted","Data":"7fe89af008642f3570a214bc910e2cd2d81c7f8a490074128ff1f03cf0048828"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.477189 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-lwhtk" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.490362 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-lwhtk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.490442 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lwhtk" podUID="4bf7a322-09ed-493a-bd28-2639f15402f5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.493740 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-xwhz4" podStartSLOduration=125.493720022 podStartE2EDuration="2m5.493720022s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.475801615 +0000 UTC m=+146.767811327" watchObservedRunningTime="2025-10-04 03:05:35.493720022 +0000 UTC m=+146.785729734" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.495530 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" event={"ID":"e5c5896d-a1f2-433c-858d-cea502c067d8","Type":"ContainerStarted","Data":"041a39e8123eaa40bd9a9588bb51e5c8f15facd86eec60a89645ddc057cd134c"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.501359 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tmzsr" podStartSLOduration=124.501317377 podStartE2EDuration="2m4.501317377s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.4932147 +0000 UTC m=+146.785224412" watchObservedRunningTime="2025-10-04 03:05:35.501317377 +0000 UTC m=+146.793327089" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.530383 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-lwhtk" podStartSLOduration=125.530360928 podStartE2EDuration="2m5.530360928s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.53004742 +0000 UTC m=+146.822057132" watchObservedRunningTime="2025-10-04 03:05:35.530360928 +0000 UTC m=+146.822370640" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.535920 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.537731 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.037708836 +0000 UTC m=+147.329718548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.567332 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" event={"ID":"d7e2aea3-e51a-42f3-a77b-1a7e8bb1e2a7","Type":"ContainerStarted","Data":"f5ca3fd475ddb9db0a3484cddd1654c802dddb26bccd4342d51e5d0c085d8aea"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.595607 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" event={"ID":"ec292e26-13cf-49b1-83a5-1c325bdd6d99","Type":"ContainerStarted","Data":"6ca14ecd792bc335c47db9840f53d8642a962456714bb7039bb872431baa2577"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.618631 4770 generic.go:334] "Generic (PLEG): container finished" podID="b153ca4e-5d53-442d-8a73-72ff80ebbe1c" containerID="5097350b75c817a5f438ae57ac1c90a77ff37b718f28f6d99e9f59bb63c2216f" exitCode=0 Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.619547 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" event={"ID":"b153ca4e-5d53-442d-8a73-72ff80ebbe1c","Type":"ContainerDied","Data":"5097350b75c817a5f438ae57ac1c90a77ff37b718f28f6d99e9f59bb63c2216f"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.637714 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.638557 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.138520411 +0000 UTC m=+147.430530123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.640452 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fbfg2" podStartSLOduration=124.640438869 podStartE2EDuration="2m4.640438869s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.634112388 +0000 UTC m=+146.926122100" watchObservedRunningTime="2025-10-04 03:05:35.640438869 +0000 UTC m=+146.932448581" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.693534 4770 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ssbgd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.693898 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" podUID="a5443b52-3d7f-4ff3-9c01-163da49074e2" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.703693 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.703729 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" event={"ID":"a5443b52-3d7f-4ff3-9c01-163da49074e2","Type":"ContainerStarted","Data":"a83056adf5078ad540bf5fa4b92e0134bf576af0dd02b9ff9d75d4af45f19ae2"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.739286 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.741932 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.241913661 +0000 UTC m=+147.533923383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.744801 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" event={"ID":"a1483904-9014-4f9d-944b-3612fbd3b3d9","Type":"ContainerStarted","Data":"f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.745990 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.758941 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" podStartSLOduration=124.758918195 podStartE2EDuration="2m4.758918195s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.6831532 +0000 UTC m=+146.975162922" watchObservedRunningTime="2025-10-04 03:05:35.758918195 +0000 UTC m=+147.050927917" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.760044 4770 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xkkrp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.760097 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" podUID="a1483904-9014-4f9d-944b-3612fbd3b3d9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.773657 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-twts6" event={"ID":"d60e800e-4ccc-4db9-ade1-4a8f43775198","Type":"ContainerStarted","Data":"4aba0373d3d4bf6c1f0482fe4624997fc86078f0f31156a6e7ed68b9b9715ed4"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.780171 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" event={"ID":"c4c445d6-f62c-4d63-8d18-35441eebeb49","Type":"ContainerStarted","Data":"f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.781304 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.786862 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" event={"ID":"61545ae7-a4d3-49dc-95a4-6393143a3718","Type":"ContainerStarted","Data":"31e14a8980837f7a1e816186a9a1f093f148426473838c6f9b87a3aae5fd1bfe"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.789108 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" podStartSLOduration=124.789096635 podStartE2EDuration="2m4.789096635s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.788188753 +0000 UTC m=+147.080198475" watchObservedRunningTime="2025-10-04 03:05:35.789096635 +0000 UTC m=+147.081106347" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.799313 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.799997 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.811540 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-v57fn" event={"ID":"0e21b999-801f-4c26-9b3f-ac4414511538","Type":"ContainerStarted","Data":"0089656f033883cd721ae7c519c9b43c66a9ea913c21e2377470f72e70458bb0"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.820426 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:35 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:35 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:35 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.820507 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.824163 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" event={"ID":"dac89ce4-6624-4d5b-a00f-5c6ffa342535","Type":"ContainerStarted","Data":"578f810e2d8d48995228fd057cb7d33cacf8f9e9d65c1de648a8da0d4fe0b5d9"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.829419 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" podStartSLOduration=125.829400765 podStartE2EDuration="2m5.829400765s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.828258575 +0000 UTC m=+147.120268287" watchObservedRunningTime="2025-10-04 03:05:35.829400765 +0000 UTC m=+147.121410477" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.840684 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.840809 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.340778935 +0000 UTC m=+147.632788647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.840993 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.852700 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.352681519 +0000 UTC m=+147.644691231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.859690 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" event={"ID":"61ef5869-de69-41b8-a0e7-5da643b24cc3","Type":"ContainerStarted","Data":"af6f5f28958e75163170ad023be55814bc9372c226c231c98b76d1023ad79620"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.868381 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" podStartSLOduration=124.868356079 podStartE2EDuration="2m4.868356079s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.867142288 +0000 UTC m=+147.159152000" watchObservedRunningTime="2025-10-04 03:05:35.868356079 +0000 UTC m=+147.160365791" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.888371 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" event={"ID":"a68c8d9f-b526-4583-901a-23595a0704a1","Type":"ContainerStarted","Data":"506f33749977b24e16de1dababc55e53faa513246f422e983101470f52c17a20"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.906340 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" podStartSLOduration=124.906310639 podStartE2EDuration="2m4.906310639s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.903832135 +0000 UTC m=+147.195841847" watchObservedRunningTime="2025-10-04 03:05:35.906310639 +0000 UTC m=+147.198320351" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.911774 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" event={"ID":"40fa4f77-7c68-4688-b6b7-234ba1dd1d33","Type":"ContainerStarted","Data":"7f0238b866d22cc94a0e18d64ceae99011aa3197d64859fed0c44a518fbbbeb2"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.913163 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.915236 4770 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-h7khz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.915284 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" podUID="40fa4f77-7c68-4688-b6b7-234ba1dd1d33" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.947399 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.947640 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-v57fn" podStartSLOduration=6.947622613 podStartE2EDuration="6.947622613s" podCreationTimestamp="2025-10-04 03:05:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:35.945182471 +0000 UTC m=+147.237192183" watchObservedRunningTime="2025-10-04 03:05:35.947622613 +0000 UTC m=+147.239632315" Oct 04 03:05:35 crc kubenswrapper[4770]: E1004 03:05:35.949744 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.449717347 +0000 UTC m=+147.741727069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.964270 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" event={"ID":"477ecb7c-f74b-4f17-87c6-e25d129aaacb","Type":"ContainerStarted","Data":"f2ea8816f7d1f8b8696c603f1e746b09ce65ae67425b7def285ddb64e0fa7fee"} Oct 04 03:05:35 crc kubenswrapper[4770]: I1004 03:05:35.974257 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" event={"ID":"e111a15d-a0c1-470b-9723-e4bbcc2265c1","Type":"ContainerStarted","Data":"1dc6b69f1c77f89981e16f02ac2ac741af3f597b1584ac24bad8a9f4eeca9384"} Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.012322 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ppsgj" podStartSLOduration=125.012289505 podStartE2EDuration="2m5.012289505s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:36.00036263 +0000 UTC m=+147.292372352" watchObservedRunningTime="2025-10-04 03:05:36.012289505 +0000 UTC m=+147.304299217" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.051577 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-pv4st" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.052159 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.062591 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.562574369 +0000 UTC m=+147.854584081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.106235 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" podStartSLOduration=125.106207883 podStartE2EDuration="2m5.106207883s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:36.087891976 +0000 UTC m=+147.379901688" watchObservedRunningTime="2025-10-04 03:05:36.106207883 +0000 UTC m=+147.398217595" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.139996 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" podStartSLOduration=125.139965005 podStartE2EDuration="2m5.139965005s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:36.138451657 +0000 UTC m=+147.430461369" watchObservedRunningTime="2025-10-04 03:05:36.139965005 +0000 UTC m=+147.431974717" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.154353 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.154848 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.654824354 +0000 UTC m=+147.946834056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.174476 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" podStartSLOduration=126.174450115 podStartE2EDuration="2m6.174450115s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:36.164469771 +0000 UTC m=+147.456479483" watchObservedRunningTime="2025-10-04 03:05:36.174450115 +0000 UTC m=+147.466459827" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.192726 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" podStartSLOduration=125.192710122 podStartE2EDuration="2m5.192710122s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:36.192303101 +0000 UTC m=+147.484312813" watchObservedRunningTime="2025-10-04 03:05:36.192710122 +0000 UTC m=+147.484719834" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.257510 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.258324 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.758305017 +0000 UTC m=+148.050314729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.358948 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.359402 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.859370618 +0000 UTC m=+148.151380330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.359560 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.359990 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.859973643 +0000 UTC m=+148.151983355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.460663 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.461122 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.961083445 +0000 UTC m=+148.253093157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.461797 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.462206 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:36.962193504 +0000 UTC m=+148.254203216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.563642 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.563845 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.063812799 +0000 UTC m=+148.355822511 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.564048 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.564407 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.064398483 +0000 UTC m=+148.356408195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.576467 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.576534 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.666001 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.666230 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.166194013 +0000 UTC m=+148.458203725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.666846 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.667536 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.167523017 +0000 UTC m=+148.459532899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.768400 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.768824 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.268806413 +0000 UTC m=+148.560816125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.805987 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:36 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:36 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:36 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.806059 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.869721 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.870154 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.37014047 +0000 UTC m=+148.662150172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.970635 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.970915 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.470879543 +0000 UTC m=+148.762889255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.970995 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:36 crc kubenswrapper[4770]: E1004 03:05:36.971349 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.471322095 +0000 UTC m=+148.763331977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.985717 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zvgls" event={"ID":"ec292e26-13cf-49b1-83a5-1c325bdd6d99","Type":"ContainerStarted","Data":"2ff0d1c4510598ed7197d3f3c766fde8b045f78c311c5ce12a299efb4ee028b2"} Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.988756 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" event={"ID":"40fa4f77-7c68-4688-b6b7-234ba1dd1d33","Type":"ContainerStarted","Data":"1f4d2648a4daf7f8dc8f0ffdd17bf85cc9433be3cacc18f31f6edbfccbd0af07"} Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.989403 4770 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-h7khz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.989436 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" podUID="40fa4f77-7c68-4688-b6b7-234ba1dd1d33" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.991264 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" event={"ID":"056a8276-f8ee-45f5-a145-63db980dc0c4","Type":"ContainerStarted","Data":"c15bb739cd26753da05961b6b76d4c33173c8920b7bf7dbfb6bb08373005b58b"} Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.992864 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-twts6" event={"ID":"d60e800e-4ccc-4db9-ade1-4a8f43775198","Type":"ContainerStarted","Data":"df7d7d29d0d6d52466f1742e3a7da95c8c0160b30e5d51730153802ba93c97b0"} Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.994774 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qq8rm" event={"ID":"477ecb7c-f74b-4f17-87c6-e25d129aaacb","Type":"ContainerStarted","Data":"42722271dab802e58021aeeac8bfeb720dcdca46043c1d5b85a5e1da4a196cec"} Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.996967 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vxv67" event={"ID":"6425e4f2-75c0-44a2-9416-76df1cb69e94","Type":"ContainerStarted","Data":"c354240619ae87101841439811a29244c7596f26e571b960a47765a1d17612fb"} Oct 04 03:05:36 crc kubenswrapper[4770]: I1004 03:05:36.998805 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" event={"ID":"0035edd0-6338-48f1-9534-3b80e135f956","Type":"ContainerStarted","Data":"66402e384299afd4ef3e5fa41f029e5c7a063f0e350205b18662bd676c18a703"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.001341 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x" event={"ID":"f54b9ac6-dfe8-4502-95c7-f657e8be4b05","Type":"ContainerStarted","Data":"5bbff05024c75fc454eb51e7f8f863a1c01b35ced73ac5a408f22a3e1d843fc7"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.002893 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" event={"ID":"a5443b52-3d7f-4ff3-9c01-163da49074e2","Type":"ContainerStarted","Data":"aa000cbf5523fe1c51521394fce9b0ff83785f50ed2a76ecfaf56689d4f5289d"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.003737 4770 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ssbgd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.003777 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" podUID="a5443b52-3d7f-4ff3-9c01-163da49074e2" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.004824 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" event={"ID":"e111a15d-a0c1-470b-9723-e4bbcc2265c1","Type":"ContainerStarted","Data":"fb77398cd17e7d7e9e65217d5b7fe4c1948ed116f2315633ca94948443af1fa2"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.006628 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" event={"ID":"dac89ce4-6624-4d5b-a00f-5c6ffa342535","Type":"ContainerStarted","Data":"ef2a0ac9704bf0d07848cc17fff8d04787b9b983797b6f601eabef4c2a58d51e"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.006839 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.008624 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x2rjg" event={"ID":"3e950818-24fa-4cbe-bdd9-1dbb895b12f1","Type":"ContainerStarted","Data":"fb829d474f5df85a032c95eaae553689ca7898338f4aeff0b8d5963f5e1bf30c"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.008655 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x2rjg" event={"ID":"3e950818-24fa-4cbe-bdd9-1dbb895b12f1","Type":"ContainerStarted","Data":"c1505eeec3aa81135719ebde0f14cd026b3c0fad1735ecc5e7433851b2451dab"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.009115 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.009202 4770 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2fqvm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.009242 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" podUID="dac89ce4-6624-4d5b-a00f-5c6ffa342535" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.011499 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" event={"ID":"941a4e7d-5fdb-4642-8d90-9a9cef86e912","Type":"ContainerStarted","Data":"ef90205bc623514e54abc84849dfa4b1a573dcab98ff6f0d1c6129cb2ecef9f1"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.011525 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" event={"ID":"941a4e7d-5fdb-4642-8d90-9a9cef86e912","Type":"ContainerStarted","Data":"0f35ce4ccaa5b8f4370543a5cbafbf15e05ff097ad7bac98a66485c7345b6f54"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.013547 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" event={"ID":"b153ca4e-5d53-442d-8a73-72ff80ebbe1c","Type":"ContainerStarted","Data":"5e0f1555380fdf9f7b7fdaa24aab2dab7cf9bd33c41db2f940185ed65e19b127"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.013935 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.015718 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" event={"ID":"b1c205f2-c2fc-49b5-8f08-7d8173506815","Type":"ContainerStarted","Data":"14baa1dae3f066969c64eaf6edef086b877f66ecbf0e77bd92436ecd66c7686a"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.017553 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jt2cp" event={"ID":"61ef5869-de69-41b8-a0e7-5da643b24cc3","Type":"ContainerStarted","Data":"39d7ad320ee45177f79068cf8df924bb7387fa7bde868189c0866333c9794193"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.019513 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" event={"ID":"e14ca55a-159f-48e5-a290-5462a4f8811c","Type":"ContainerStarted","Data":"20b11cc6633ad17b3ab70eb1f909873a33f9f7e4e8f9f2b121ade67533fa3a4e"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.021122 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" event={"ID":"e9d25eff-feac-415e-9ba2-e34de3c0d866","Type":"ContainerStarted","Data":"d4cc9e3a4e26f29fcb75cf7cd73ff8fd6df586fd93322a5a76115e208f913732"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.023079 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" event={"ID":"61545ae7-a4d3-49dc-95a4-6393143a3718","Type":"ContainerStarted","Data":"40f7667cdef34b5cb947e61d9ed196812ab5c3c19cac917693a057b1884b1769"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.025734 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" event={"ID":"7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb","Type":"ContainerStarted","Data":"8019dc3c35e37bae7f47a0e21bb51c5356a9f1bb8fde94e53baa1ad680ec7924"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.025772 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.025788 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" event={"ID":"7b8bd9ee-9b48-462f-9a2e-83a996b9d5eb","Type":"ContainerStarted","Data":"9f21042be7648c352e72494b7714975fdb49c83cb1d496e4d252ceae02afb627"} Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.030258 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-lwhtk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.030327 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lwhtk" podUID="4bf7a322-09ed-493a-bd28-2639f15402f5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.044765 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.072583 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.074652 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.574613332 +0000 UTC m=+148.866623054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.108884 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.137948 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8srq6" podStartSLOduration=126.137917409 podStartE2EDuration="2m6.137917409s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.059472476 +0000 UTC m=+148.351482188" watchObservedRunningTime="2025-10-04 03:05:37.137917409 +0000 UTC m=+148.429927121" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.138426 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rgdpf" podStartSLOduration=126.138420572 podStartE2EDuration="2m6.138420572s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.113733531 +0000 UTC m=+148.405743243" watchObservedRunningTime="2025-10-04 03:05:37.138420572 +0000 UTC m=+148.430430284" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.163429 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s26vf" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.176580 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.177198 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.677175832 +0000 UTC m=+148.969185534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.241729 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" podStartSLOduration=126.241707119 podStartE2EDuration="2m6.241707119s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.223498704 +0000 UTC m=+148.515508416" watchObservedRunningTime="2025-10-04 03:05:37.241707119 +0000 UTC m=+148.533716831" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.278588 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.279163 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.779140205 +0000 UTC m=+149.071149917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.380483 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.381061 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.881044577 +0000 UTC m=+149.173054279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.481788 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.481988 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.981945393 +0000 UTC m=+149.273955105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.482257 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.482797 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:37.982776315 +0000 UTC m=+149.274786027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.488346 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" podStartSLOduration=126.488321847 podStartE2EDuration="2m6.488321847s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.346205457 +0000 UTC m=+148.638215169" watchObservedRunningTime="2025-10-04 03:05:37.488321847 +0000 UTC m=+148.780331559" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.550373 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-twts6" podStartSLOduration=8.5503503 podStartE2EDuration="8.5503503s" podCreationTimestamp="2025-10-04 03:05:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.546986435 +0000 UTC m=+148.838996147" watchObservedRunningTime="2025-10-04 03:05:37.5503503 +0000 UTC m=+148.842360012" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.550705 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-x2rjg" podStartSLOduration=8.550700279 podStartE2EDuration="8.550700279s" podCreationTimestamp="2025-10-04 03:05:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.512362521 +0000 UTC m=+148.804372253" watchObservedRunningTime="2025-10-04 03:05:37.550700279 +0000 UTC m=+148.842709991" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.583773 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.584188 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.084168264 +0000 UTC m=+149.376177976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.639442 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ntlrg" podStartSLOduration=126.63941968500001 podStartE2EDuration="2m6.639419685s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.638949723 +0000 UTC m=+148.930959445" watchObservedRunningTime="2025-10-04 03:05:37.639419685 +0000 UTC m=+148.931429397" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.641264 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-nmsxt" podStartSLOduration=126.641237461 podStartE2EDuration="2m6.641237461s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.59455472 +0000 UTC m=+148.886564432" watchObservedRunningTime="2025-10-04 03:05:37.641237461 +0000 UTC m=+148.933247183" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.666793 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.685385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.686026 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.185972814 +0000 UTC m=+149.477982526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.717540 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" podStartSLOduration=126.71751567 podStartE2EDuration="2m6.71751567s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.715135219 +0000 UTC m=+149.007144931" watchObservedRunningTime="2025-10-04 03:05:37.71751567 +0000 UTC m=+149.009525382" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.740482 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-kpt6s" podStartSLOduration=126.740458205 podStartE2EDuration="2m6.740458205s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.735681443 +0000 UTC m=+149.027691155" watchObservedRunningTime="2025-10-04 03:05:37.740458205 +0000 UTC m=+149.032467917" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.787405 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.787798 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.287781214 +0000 UTC m=+149.579790926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.811200 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:37 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:37 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:37 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.811269 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.889239 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.889814 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.389774609 +0000 UTC m=+149.681784321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.894132 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bzk7x" podStartSLOduration=126.894118659 podStartE2EDuration="2m6.894118659s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.879952648 +0000 UTC m=+149.171962370" watchObservedRunningTime="2025-10-04 03:05:37.894118659 +0000 UTC m=+149.186128371" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.990072 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.990296 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.490262824 +0000 UTC m=+149.782272536 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.990539 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:37 crc kubenswrapper[4770]: E1004 03:05:37.990916 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.490904411 +0000 UTC m=+149.782914123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.997578 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" podStartSLOduration=127.997554141 podStartE2EDuration="2m7.997554141s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.932584471 +0000 UTC m=+149.224594183" watchObservedRunningTime="2025-10-04 03:05:37.997554141 +0000 UTC m=+149.289563853" Oct 04 03:05:37 crc kubenswrapper[4770]: I1004 03:05:37.997874 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fxv82" podStartSLOduration=126.997870709 podStartE2EDuration="2m6.997870709s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:37.989002931 +0000 UTC m=+149.281012643" watchObservedRunningTime="2025-10-04 03:05:37.997870709 +0000 UTC m=+149.289880421" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.032954 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-lwhtk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.033083 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lwhtk" podUID="4bf7a322-09ed-493a-bd28-2639f15402f5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.033736 4770 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2fqvm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.033787 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" podUID="dac89ce4-6624-4d5b-a00f-5c6ffa342535" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.042871 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5m7kk" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.047184 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h7khz" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.091932 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:38 crc kubenswrapper[4770]: E1004 03:05:38.097870 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.597831731 +0000 UTC m=+149.889841443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.198684 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:38 crc kubenswrapper[4770]: E1004 03:05:38.199191 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.699178249 +0000 UTC m=+149.991187961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.303299 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:38 crc kubenswrapper[4770]: E1004 03:05:38.303837 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.803815961 +0000 UTC m=+150.095825673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.408888 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:38 crc kubenswrapper[4770]: E1004 03:05:38.409383 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:38.909360996 +0000 UTC m=+150.201370708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.510848 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.511290 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.511332 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.511354 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.511379 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.515091 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:38 crc kubenswrapper[4770]: E1004 03:05:38.515757 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.015725313 +0000 UTC m=+150.307735025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.528150 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.532153 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.536783 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.544785 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dcs7t"] Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.546084 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.568673 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.589363 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dcs7t"] Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.594419 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.614927 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4fdc\" (UniqueName: \"kubernetes.io/projected/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-kube-api-access-l4fdc\") pod \"community-operators-dcs7t\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.615193 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.615321 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-catalog-content\") pod \"community-operators-dcs7t\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.615410 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-utilities\") pod \"community-operators-dcs7t\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: E1004 03:05:38.615648 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.115627293 +0000 UTC m=+150.407637005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.700384 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.708222 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.720579 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.720749 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-catalog-content\") pod \"community-operators-dcs7t\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.720780 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-utilities\") pod \"community-operators-dcs7t\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.720846 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4fdc\" (UniqueName: \"kubernetes.io/projected/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-kube-api-access-l4fdc\") pod \"community-operators-dcs7t\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.721484 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-catalog-content\") pod \"community-operators-dcs7t\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.721583 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-utilities\") pod \"community-operators-dcs7t\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: E1004 03:05:38.722175 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.222138963 +0000 UTC m=+150.514148675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.776157 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4fdc\" (UniqueName: \"kubernetes.io/projected/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-kube-api-access-l4fdc\") pod \"community-operators-dcs7t\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.808278 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:38 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:38 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:38 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.808876 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.824051 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:38 crc kubenswrapper[4770]: E1004 03:05:38.824506 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.324491097 +0000 UTC m=+150.616500799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.874366 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-97vj5"] Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.875501 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.883992 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.899788 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-97vj5"] Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.924973 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.925273 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtfg6\" (UniqueName: \"kubernetes.io/projected/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-kube-api-access-wtfg6\") pod \"community-operators-97vj5\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.925321 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-utilities\") pod \"community-operators-97vj5\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:38 crc kubenswrapper[4770]: I1004 03:05:38.925358 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-catalog-content\") pod \"community-operators-97vj5\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:38 crc kubenswrapper[4770]: E1004 03:05:38.925568 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.425532317 +0000 UTC m=+150.717542029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.030326 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtfg6\" (UniqueName: \"kubernetes.io/projected/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-kube-api-access-wtfg6\") pod \"community-operators-97vj5\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.030379 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-utilities\") pod \"community-operators-97vj5\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.030411 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-catalog-content\") pod \"community-operators-97vj5\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.030456 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.031070 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.531056242 +0000 UTC m=+150.823065954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.031440 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-catalog-content\") pod \"community-operators-97vj5\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.031535 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-utilities\") pod \"community-operators-97vj5\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.032131 4770 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ssbgd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.032174 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" podUID="a5443b52-3d7f-4ff3-9c01-163da49074e2" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.084558 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtfg6\" (UniqueName: \"kubernetes.io/projected/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-kube-api-access-wtfg6\") pod \"community-operators-97vj5\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.090128 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vxv67" event={"ID":"6425e4f2-75c0-44a2-9416-76df1cb69e94","Type":"ContainerStarted","Data":"30eda616a24fa9e4dfbae4e87bb151b6ec784c7a0545ccbd6611e54efaa32f65"} Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.090212 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m4jht"] Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.092853 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vxv67" event={"ID":"6425e4f2-75c0-44a2-9416-76df1cb69e94","Type":"ContainerStarted","Data":"c2d43c3973f38aeabfbf42c51dd32d6a8e061010586d7dc8711cf7ca09126be1"} Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.092940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.104653 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.132593 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.132870 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-utilities\") pod \"certified-operators-m4jht\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.132937 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-catalog-content\") pod \"certified-operators-m4jht\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.133137 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s28jc\" (UniqueName: \"kubernetes.io/projected/d6d87160-c0c6-4cb0-8561-13988e9c5395-kube-api-access-s28jc\") pod \"certified-operators-m4jht\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.134491 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.634461013 +0000 UTC m=+150.926470735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.200379 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.222673 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m4jht"] Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.235403 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.235492 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-utilities\") pod \"certified-operators-m4jht\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.235521 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-catalog-content\") pod \"certified-operators-m4jht\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.235571 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s28jc\" (UniqueName: \"kubernetes.io/projected/d6d87160-c0c6-4cb0-8561-13988e9c5395-kube-api-access-s28jc\") pod \"certified-operators-m4jht\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.236346 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.736333323 +0000 UTC m=+151.028343035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.237091 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-utilities\") pod \"certified-operators-m4jht\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.253116 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-catalog-content\") pod \"certified-operators-m4jht\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.277063 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s28jc\" (UniqueName: \"kubernetes.io/projected/d6d87160-c0c6-4cb0-8561-13988e9c5395-kube-api-access-s28jc\") pod \"certified-operators-m4jht\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.317164 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rnkr7"] Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.318565 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.331978 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rnkr7"] Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.340683 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.341541 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.841519489 +0000 UTC m=+151.133529191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.444976 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-utilities\") pod \"certified-operators-rnkr7\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.445068 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v5g9\" (UniqueName: \"kubernetes.io/projected/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-kube-api-access-9v5g9\") pod \"certified-operators-rnkr7\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.445120 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.445156 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-catalog-content\") pod \"certified-operators-rnkr7\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.446040 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:39.946025458 +0000 UTC m=+151.238035170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.508577 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.548680 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.550482 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:40.050445955 +0000 UTC m=+151.342455667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.548997 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-catalog-content\") pod \"certified-operators-rnkr7\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.562417 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-utilities\") pod \"certified-operators-rnkr7\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.562522 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v5g9\" (UniqueName: \"kubernetes.io/projected/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-kube-api-access-9v5g9\") pod \"certified-operators-rnkr7\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.566483 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-catalog-content\") pod \"certified-operators-rnkr7\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.568371 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-utilities\") pod \"certified-operators-rnkr7\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.628909 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v5g9\" (UniqueName: \"kubernetes.io/projected/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-kube-api-access-9v5g9\") pod \"certified-operators-rnkr7\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.664176 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.664595 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:40.164579409 +0000 UTC m=+151.456589121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.688472 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.770925 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.771446 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:40.271424507 +0000 UTC m=+151.563434219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.813086 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dcs7t"] Oct 04 03:05:39 crc kubenswrapper[4770]: W1004 03:05:39.825170 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-89034d96046245485816554c06f2d7e30f5b675e8a04e9a4b001ac54461c18ee WatchSource:0}: Error finding container 89034d96046245485816554c06f2d7e30f5b675e8a04e9a4b001ac54461c18ee: Status 404 returned error can't find the container with id 89034d96046245485816554c06f2d7e30f5b675e8a04e9a4b001ac54461c18ee Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.825264 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:39 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:39 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:39 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.825306 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.854594 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lt6hf" Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.874333 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.874743 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:40.374728845 +0000 UTC m=+151.666738557 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:39 crc kubenswrapper[4770]: I1004 03:05:39.976709 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:39 crc kubenswrapper[4770]: E1004 03:05:39.977043 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:40.477023297 +0000 UTC m=+151.769033009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.031033 4770 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.080525 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.080984 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:40.580965472 +0000 UTC m=+151.872975184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.117247 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-97vj5"] Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.184622 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.185168 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:40.685145472 +0000 UTC m=+151.977155184 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.185457 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vxv67" event={"ID":"6425e4f2-75c0-44a2-9416-76df1cb69e94","Type":"ContainerStarted","Data":"43b20fdaf23d6723f383413561652b7960a32053f8675907c4f8a6cbc98e8230"} Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.235681 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4b52d7813b8ed5bce9bb7093b81c368eb909d966d4a578c1ca5fcc2e162767ef"} Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.258655 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"89034d96046245485816554c06f2d7e30f5b675e8a04e9a4b001ac54461c18ee"} Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.297157 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.297480 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:40.79746628 +0000 UTC m=+152.089475992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.304915 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7c7a9813f998b70bd793095d3340bbe5653e6abddaff146446608efabbd555a5"} Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.308609 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dcs7t" event={"ID":"30e80d37-aeb7-4446-852c-3a3ac2b78bb8","Type":"ContainerStarted","Data":"592586b37ed5f2c9bdb81150ff5c9f5f71a844ad5075c6e3af33b0b61af59679"} Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.402640 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.404158 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:40.904136364 +0000 UTC m=+152.196146076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.463292 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-vxv67" podStartSLOduration=11.463274255 podStartE2EDuration="11.463274255s" podCreationTimestamp="2025-10-04 03:05:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:40.265297379 +0000 UTC m=+151.557307091" watchObservedRunningTime="2025-10-04 03:05:40.463274255 +0000 UTC m=+151.755283967" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.467549 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m4jht"] Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.503737 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.504089 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:41.004076117 +0000 UTC m=+152.296085829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.604570 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.604688 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:41.104666446 +0000 UTC m=+152.396676158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.605105 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.605607 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:41.105585418 +0000 UTC m=+152.397595130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.624418 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rnkr7"] Oct 04 03:05:40 crc kubenswrapper[4770]: W1004 03:05:40.653716 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9dbdce0_d5c5_4b60_9d4b_745beca12ee8.slice/crio-f5f3b765ec21011bd2c2374ad7383a5790fc383cda592613efd6c8844b3ab89a WatchSource:0}: Error finding container f5f3b765ec21011bd2c2374ad7383a5790fc383cda592613efd6c8844b3ab89a: Status 404 returned error can't find the container with id f5f3b765ec21011bd2c2374ad7383a5790fc383cda592613efd6c8844b3ab89a Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.658968 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n9zkn"] Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.662373 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.665236 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.673359 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9zkn"] Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.679675 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.680723 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.684558 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.693418 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.699212 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30e80d37_aeb7_4446_852c_3a3ac2b78bb8.slice/crio-fc49be386e3d7e999e25a6a07b2924343cbffcd4cc36045c228be0f0f096720d.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.701288 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.713808 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.714069 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:41.214032048 +0000 UTC m=+152.506041760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.804375 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:40 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:40 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:40 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.804435 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.815540 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp6xb\" (UniqueName: \"kubernetes.io/projected/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-kube-api-access-hp6xb\") pod \"redhat-marketplace-n9zkn\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.815601 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-catalog-content\") pod \"redhat-marketplace-n9zkn\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.815637 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.815663 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-utilities\") pod \"redhat-marketplace-n9zkn\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.815897 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/517c9323-8c40-413a-92a8-9dd833dbfc2c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"517c9323-8c40-413a-92a8-9dd833dbfc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.816097 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/517c9323-8c40-413a-92a8-9dd833dbfc2c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"517c9323-8c40-413a-92a8-9dd833dbfc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.817910 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:41.31786399 +0000 UTC m=+152.609873702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.919162 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.919321 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp6xb\" (UniqueName: \"kubernetes.io/projected/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-kube-api-access-hp6xb\") pod \"redhat-marketplace-n9zkn\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.919397 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:05:41.41934528 +0000 UTC m=+152.711354992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.919497 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-catalog-content\") pod \"redhat-marketplace-n9zkn\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.919683 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-utilities\") pod \"redhat-marketplace-n9zkn\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.919737 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.919814 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/517c9323-8c40-413a-92a8-9dd833dbfc2c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"517c9323-8c40-413a-92a8-9dd833dbfc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.919964 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/517c9323-8c40-413a-92a8-9dd833dbfc2c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"517c9323-8c40-413a-92a8-9dd833dbfc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.920847 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-catalog-content\") pod \"redhat-marketplace-n9zkn\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:40 crc kubenswrapper[4770]: E1004 03:05:40.920926 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:05:41.42091395 +0000 UTC m=+152.712923662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-lnhsd" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.920927 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-utilities\") pod \"redhat-marketplace-n9zkn\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.920975 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/517c9323-8c40-413a-92a8-9dd833dbfc2c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"517c9323-8c40-413a-92a8-9dd833dbfc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.926663 4770 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-04T03:05:40.043177657Z","Handler":null,"Name":""} Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.935098 4770 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.935160 4770 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.942644 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/517c9323-8c40-413a-92a8-9dd833dbfc2c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"517c9323-8c40-413a-92a8-9dd833dbfc2c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:40 crc kubenswrapper[4770]: I1004 03:05:40.946836 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp6xb\" (UniqueName: \"kubernetes.io/projected/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-kube-api-access-hp6xb\") pod \"redhat-marketplace-n9zkn\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.020843 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.030532 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.039465 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.061273 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.068461 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bk4rd"] Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.069789 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.085694 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk4rd"] Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.123607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.130656 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.130696 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.179404 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-lnhsd\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.225197 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pmd4\" (UniqueName: \"kubernetes.io/projected/400cc27a-45ec-4164-9af6-7c120fff6351-kube-api-access-4pmd4\") pod \"redhat-marketplace-bk4rd\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.225271 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-utilities\") pod \"redhat-marketplace-bk4rd\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.225353 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-catalog-content\") pod \"redhat-marketplace-bk4rd\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.262773 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.298830 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9zkn"] Oct 04 03:05:41 crc kubenswrapper[4770]: W1004 03:05:41.319651 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbba2dc2c_e45e_43c2_8d02_ee01bf41143a.slice/crio-7f8808ef8da203448f17c401a8e45dd322d75438cd763b08cc00cd2bbf492a89 WatchSource:0}: Error finding container 7f8808ef8da203448f17c401a8e45dd322d75438cd763b08cc00cd2bbf492a89: Status 404 returned error can't find the container with id 7f8808ef8da203448f17c401a8e45dd322d75438cd763b08cc00cd2bbf492a89 Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.324563 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b47832473860629750fcb3dbb86a464200e9148b49aa9871f39c2838ae7cb0f6"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.324797 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.327832 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-catalog-content\") pod \"redhat-marketplace-bk4rd\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.327888 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pmd4\" (UniqueName: \"kubernetes.io/projected/400cc27a-45ec-4164-9af6-7c120fff6351-kube-api-access-4pmd4\") pod \"redhat-marketplace-bk4rd\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.327916 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-utilities\") pod \"redhat-marketplace-bk4rd\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.328341 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-utilities\") pod \"redhat-marketplace-bk4rd\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.328596 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-catalog-content\") pod \"redhat-marketplace-bk4rd\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.329771 4770 generic.go:334] "Generic (PLEG): container finished" podID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerID="fc49be386e3d7e999e25a6a07b2924343cbffcd4cc36045c228be0f0f096720d" exitCode=0 Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.329923 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dcs7t" event={"ID":"30e80d37-aeb7-4446-852c-3a3ac2b78bb8","Type":"ContainerDied","Data":"fc49be386e3d7e999e25a6a07b2924343cbffcd4cc36045c228be0f0f096720d"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.336120 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"abc800d7d56cd37df7180a5e3a8677e493ace678bc9efb277e0ff81af88ccb4f"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.349709 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.352497 4770 generic.go:334] "Generic (PLEG): container finished" podID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerID="a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3" exitCode=0 Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.352588 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rnkr7" event={"ID":"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8","Type":"ContainerDied","Data":"a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.352638 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rnkr7" event={"ID":"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8","Type":"ContainerStarted","Data":"f5f3b765ec21011bd2c2374ad7383a5790fc383cda592613efd6c8844b3ab89a"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.359948 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pmd4\" (UniqueName: \"kubernetes.io/projected/400cc27a-45ec-4164-9af6-7c120fff6351-kube-api-access-4pmd4\") pod \"redhat-marketplace-bk4rd\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.387501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"28629dbec5bbbf715d754cd7a5efe0ccd0310f5689ed7c8069cfbdfa5f29e771"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.393725 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.409794 4770 generic.go:334] "Generic (PLEG): container finished" podID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerID="1d6ed5cbe64527c2ef27923e626e083919ee17042ce5caa5a31dbfc1a71b3d60" exitCode=0 Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.409876 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4jht" event={"ID":"d6d87160-c0c6-4cb0-8561-13988e9c5395","Type":"ContainerDied","Data":"1d6ed5cbe64527c2ef27923e626e083919ee17042ce5caa5a31dbfc1a71b3d60"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.409910 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4jht" event={"ID":"d6d87160-c0c6-4cb0-8561-13988e9c5395","Type":"ContainerStarted","Data":"d9628b2898b3c0db9cf5615a222b4ddd63da2f3c1c360b038e95ee508489ddb6"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.420285 4770 generic.go:334] "Generic (PLEG): container finished" podID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerID="47f4e67aa2ca7bcf80ffe0f1a8f1cac3afe778c54efd21f5432355bb52b0146f" exitCode=0 Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.421152 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97vj5" event={"ID":"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1","Type":"ContainerDied","Data":"47f4e67aa2ca7bcf80ffe0f1a8f1cac3afe778c54efd21f5432355bb52b0146f"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.421184 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97vj5" event={"ID":"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1","Type":"ContainerStarted","Data":"f7b8e2ceaa4dec5684455baed16bc1a1cd055d5bad952c94b7d2f0b653b1e59b"} Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.440705 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.608837 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lnhsd"] Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.690282 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.773672 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk4rd"] Oct 04 03:05:41 crc kubenswrapper[4770]: W1004 03:05:41.790697 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod400cc27a_45ec_4164_9af6_7c120fff6351.slice/crio-1ecdb793ad527a3635d4c0b97107475f94e0d1938c9e063cca2985424c4098d6 WatchSource:0}: Error finding container 1ecdb793ad527a3635d4c0b97107475f94e0d1938c9e063cca2985424c4098d6: Status 404 returned error can't find the container with id 1ecdb793ad527a3635d4c0b97107475f94e0d1938c9e063cca2985424c4098d6 Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.804923 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:41 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:41 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:41 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.805032 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.943030 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.943081 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:41 crc kubenswrapper[4770]: I1004 03:05:41.951281 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.064429 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pvjjj"] Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.066214 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.070485 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.128823 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pvjjj"] Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.148120 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-catalog-content\") pod \"redhat-operators-pvjjj\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.148295 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jzwc\" (UniqueName: \"kubernetes.io/projected/f8207392-9683-4278-a7d5-8e6af62fbf61-kube-api-access-7jzwc\") pod \"redhat-operators-pvjjj\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.148325 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-utilities\") pod \"redhat-operators-pvjjj\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.251249 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jzwc\" (UniqueName: \"kubernetes.io/projected/f8207392-9683-4278-a7d5-8e6af62fbf61-kube-api-access-7jzwc\") pod \"redhat-operators-pvjjj\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.251309 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-utilities\") pod \"redhat-operators-pvjjj\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.251376 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-catalog-content\") pod \"redhat-operators-pvjjj\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.252154 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-catalog-content\") pod \"redhat-operators-pvjjj\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.252205 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-utilities\") pod \"redhat-operators-pvjjj\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.289362 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jzwc\" (UniqueName: \"kubernetes.io/projected/f8207392-9683-4278-a7d5-8e6af62fbf61-kube-api-access-7jzwc\") pod \"redhat-operators-pvjjj\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.369299 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.369371 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.388521 4770 patch_prober.go:28] interesting pod/console-f9d7485db-xwhz4 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.388597 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-xwhz4" podUID="2b45833c-2e11-428e-8131-1ea4676a794f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.388750 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-lwhtk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.388806 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lwhtk" podUID="4bf7a322-09ed-493a-bd28-2639f15402f5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.388918 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-lwhtk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.388977 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lwhtk" podUID="4bf7a322-09ed-493a-bd28-2639f15402f5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.391353 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.439702 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"517c9323-8c40-413a-92a8-9dd833dbfc2c","Type":"ContainerStarted","Data":"9bb5a470405663001f7dd0e1f4c844e8e8a4278e5d3ec1aa8ac5248058a7d790"} Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.439752 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"517c9323-8c40-413a-92a8-9dd833dbfc2c","Type":"ContainerStarted","Data":"5dad8233878c85ad282f4541b0c29daeabe08626f2672bda566b907769415240"} Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.451934 4770 generic.go:334] "Generic (PLEG): container finished" podID="400cc27a-45ec-4164-9af6-7c120fff6351" containerID="df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0" exitCode=0 Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.452064 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk4rd" event={"ID":"400cc27a-45ec-4164-9af6-7c120fff6351","Type":"ContainerDied","Data":"df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0"} Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.452141 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk4rd" event={"ID":"400cc27a-45ec-4164-9af6-7c120fff6351","Type":"ContainerStarted","Data":"1ecdb793ad527a3635d4c0b97107475f94e0d1938c9e063cca2985424c4098d6"} Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.466109 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zzdjv"] Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.467587 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.473435 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.473408225 podStartE2EDuration="2.473408225s" podCreationTimestamp="2025-10-04 03:05:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:42.471737502 +0000 UTC m=+153.763747214" watchObservedRunningTime="2025-10-04 03:05:42.473408225 +0000 UTC m=+153.765417937" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.481704 4770 generic.go:334] "Generic (PLEG): container finished" podID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerID="5cc3549c07c2da4cb528206b1e762e2847d1987f979f0546c6c83a820b78fb1a" exitCode=0 Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.481790 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9zkn" event={"ID":"bba2dc2c-e45e-43c2-8d02-ee01bf41143a","Type":"ContainerDied","Data":"5cc3549c07c2da4cb528206b1e762e2847d1987f979f0546c6c83a820b78fb1a"} Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.481824 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9zkn" event={"ID":"bba2dc2c-e45e-43c2-8d02-ee01bf41143a","Type":"ContainerStarted","Data":"7f8808ef8da203448f17c401a8e45dd322d75438cd763b08cc00cd2bbf492a89"} Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.491470 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzdjv"] Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.501390 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" event={"ID":"108c220a-bc76-4e17-884d-80f2a5b63357","Type":"ContainerStarted","Data":"72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81"} Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.501434 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.501446 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" event={"ID":"108c220a-bc76-4e17-884d-80f2a5b63357","Type":"ContainerStarted","Data":"ac09181def512480c2128ccbeea7dd6d5bebc64b6e45f6196bb5ccb1b44efadc"} Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.507785 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-pnsm5" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.555295 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-catalog-content\") pod \"redhat-operators-zzdjv\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.555863 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7skx\" (UniqueName: \"kubernetes.io/projected/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-kube-api-access-w7skx\") pod \"redhat-operators-zzdjv\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.555896 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-utilities\") pod \"redhat-operators-zzdjv\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.585126 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" podStartSLOduration=131.585109638 podStartE2EDuration="2m11.585109638s" podCreationTimestamp="2025-10-04 03:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:05:42.56913568 +0000 UTC m=+153.861145402" watchObservedRunningTime="2025-10-04 03:05:42.585109638 +0000 UTC m=+153.877119350" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.658889 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-catalog-content\") pod \"redhat-operators-zzdjv\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.658983 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7skx\" (UniqueName: \"kubernetes.io/projected/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-kube-api-access-w7skx\") pod \"redhat-operators-zzdjv\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.659033 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-utilities\") pod \"redhat-operators-zzdjv\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.680789 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-utilities\") pod \"redhat-operators-zzdjv\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.690878 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-catalog-content\") pod \"redhat-operators-zzdjv\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.714221 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7skx\" (UniqueName: \"kubernetes.io/projected/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-kube-api-access-w7skx\") pod \"redhat-operators-zzdjv\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.801444 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.810535 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:42 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:42 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:42 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.810610 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.834952 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:05:42 crc kubenswrapper[4770]: I1004 03:05:42.856806 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ssbgd" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.069831 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pvjjj"] Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.195738 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.197714 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.197821 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.203799 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.204046 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.206907 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.278607 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e54166e-b9e8-4136-9f06-09e8480555b7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4e54166e-b9e8-4136-9f06-09e8480555b7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.278659 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e54166e-b9e8-4136-9f06-09e8480555b7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4e54166e-b9e8-4136-9f06-09e8480555b7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.379975 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e54166e-b9e8-4136-9f06-09e8480555b7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4e54166e-b9e8-4136-9f06-09e8480555b7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.380391 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e54166e-b9e8-4136-9f06-09e8480555b7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4e54166e-b9e8-4136-9f06-09e8480555b7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.380697 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e54166e-b9e8-4136-9f06-09e8480555b7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4e54166e-b9e8-4136-9f06-09e8480555b7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.400381 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e54166e-b9e8-4136-9f06-09e8480555b7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4e54166e-b9e8-4136-9f06-09e8480555b7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.531797 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.557519 4770 generic.go:334] "Generic (PLEG): container finished" podID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerID="307131fe5936053720116f32bfdc419f2ffd674a57db8efb2fe235020c3f0055" exitCode=0 Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.557624 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjjj" event={"ID":"f8207392-9683-4278-a7d5-8e6af62fbf61","Type":"ContainerDied","Data":"307131fe5936053720116f32bfdc419f2ffd674a57db8efb2fe235020c3f0055"} Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.557666 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjjj" event={"ID":"f8207392-9683-4278-a7d5-8e6af62fbf61","Type":"ContainerStarted","Data":"0f3fe2e8f4f45d4fc5d1974e16ae7d11236f65cdbf5818d50d234286538a4a34"} Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.564806 4770 generic.go:334] "Generic (PLEG): container finished" podID="517c9323-8c40-413a-92a8-9dd833dbfc2c" containerID="9bb5a470405663001f7dd0e1f4c844e8e8a4278e5d3ec1aa8ac5248058a7d790" exitCode=0 Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.564905 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"517c9323-8c40-413a-92a8-9dd833dbfc2c","Type":"ContainerDied","Data":"9bb5a470405663001f7dd0e1f4c844e8e8a4278e5d3ec1aa8ac5248058a7d790"} Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.568814 4770 generic.go:334] "Generic (PLEG): container finished" podID="e111a15d-a0c1-470b-9723-e4bbcc2265c1" containerID="fb77398cd17e7d7e9e65217d5b7fe4c1948ed116f2315633ca94948443af1fa2" exitCode=0 Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.569000 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" event={"ID":"e111a15d-a0c1-470b-9723-e4bbcc2265c1","Type":"ContainerDied","Data":"fb77398cd17e7d7e9e65217d5b7fe4c1948ed116f2315633ca94948443af1fa2"} Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.648066 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzdjv"] Oct 04 03:05:43 crc kubenswrapper[4770]: W1004 03:05:43.680343 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac60a2ae_8790_4e3d_a2c0_ba394a8140a6.slice/crio-2f95f0890cb1e2708fb936668e408ea77360245e1a2e2fa43c7a4dde8dc22927 WatchSource:0}: Error finding container 2f95f0890cb1e2708fb936668e408ea77360245e1a2e2fa43c7a4dde8dc22927: Status 404 returned error can't find the container with id 2f95f0890cb1e2708fb936668e408ea77360245e1a2e2fa43c7a4dde8dc22927 Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.806309 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:43 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:43 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:43 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:43 crc kubenswrapper[4770]: I1004 03:05:43.806391 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:44 crc kubenswrapper[4770]: I1004 03:05:44.217151 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 03:05:44 crc kubenswrapper[4770]: W1004 03:05:44.236412 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4e54166e_b9e8_4136_9f06_09e8480555b7.slice/crio-dbe5363c01869ff218ebf97d58173e70524043ace496a13b7f5f5e634be228ea WatchSource:0}: Error finding container dbe5363c01869ff218ebf97d58173e70524043ace496a13b7f5f5e634be228ea: Status 404 returned error can't find the container with id dbe5363c01869ff218ebf97d58173e70524043ace496a13b7f5f5e634be228ea Oct 04 03:05:44 crc kubenswrapper[4770]: I1004 03:05:44.583115 4770 generic.go:334] "Generic (PLEG): container finished" podID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerID="cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2" exitCode=0 Oct 04 03:05:44 crc kubenswrapper[4770]: I1004 03:05:44.583216 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzdjv" event={"ID":"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6","Type":"ContainerDied","Data":"cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2"} Oct 04 03:05:44 crc kubenswrapper[4770]: I1004 03:05:44.583680 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzdjv" event={"ID":"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6","Type":"ContainerStarted","Data":"2f95f0890cb1e2708fb936668e408ea77360245e1a2e2fa43c7a4dde8dc22927"} Oct 04 03:05:44 crc kubenswrapper[4770]: I1004 03:05:44.586452 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4e54166e-b9e8-4136-9f06-09e8480555b7","Type":"ContainerStarted","Data":"dbe5363c01869ff218ebf97d58173e70524043ace496a13b7f5f5e634be228ea"} Oct 04 03:05:44 crc kubenswrapper[4770]: I1004 03:05:44.809358 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:44 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:44 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:44 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:44 crc kubenswrapper[4770]: I1004 03:05:44.809436 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:44 crc kubenswrapper[4770]: I1004 03:05:44.921516 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:44 crc kubenswrapper[4770]: I1004 03:05:44.969076 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.013721 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cztnz\" (UniqueName: \"kubernetes.io/projected/e111a15d-a0c1-470b-9723-e4bbcc2265c1-kube-api-access-cztnz\") pod \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.013801 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e111a15d-a0c1-470b-9723-e4bbcc2265c1-config-volume\") pod \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.013931 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e111a15d-a0c1-470b-9723-e4bbcc2265c1-secret-volume\") pod \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\" (UID: \"e111a15d-a0c1-470b-9723-e4bbcc2265c1\") " Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.015083 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e111a15d-a0c1-470b-9723-e4bbcc2265c1-config-volume" (OuterVolumeSpecName: "config-volume") pod "e111a15d-a0c1-470b-9723-e4bbcc2265c1" (UID: "e111a15d-a0c1-470b-9723-e4bbcc2265c1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.024281 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e111a15d-a0c1-470b-9723-e4bbcc2265c1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e111a15d-a0c1-470b-9723-e4bbcc2265c1" (UID: "e111a15d-a0c1-470b-9723-e4bbcc2265c1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.027191 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e111a15d-a0c1-470b-9723-e4bbcc2265c1-kube-api-access-cztnz" (OuterVolumeSpecName: "kube-api-access-cztnz") pod "e111a15d-a0c1-470b-9723-e4bbcc2265c1" (UID: "e111a15d-a0c1-470b-9723-e4bbcc2265c1"). InnerVolumeSpecName "kube-api-access-cztnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.116118 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/517c9323-8c40-413a-92a8-9dd833dbfc2c-kube-api-access\") pod \"517c9323-8c40-413a-92a8-9dd833dbfc2c\" (UID: \"517c9323-8c40-413a-92a8-9dd833dbfc2c\") " Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.117594 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/517c9323-8c40-413a-92a8-9dd833dbfc2c-kubelet-dir\") pod \"517c9323-8c40-413a-92a8-9dd833dbfc2c\" (UID: \"517c9323-8c40-413a-92a8-9dd833dbfc2c\") " Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.117919 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/517c9323-8c40-413a-92a8-9dd833dbfc2c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "517c9323-8c40-413a-92a8-9dd833dbfc2c" (UID: "517c9323-8c40-413a-92a8-9dd833dbfc2c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.118281 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cztnz\" (UniqueName: \"kubernetes.io/projected/e111a15d-a0c1-470b-9723-e4bbcc2265c1-kube-api-access-cztnz\") on node \"crc\" DevicePath \"\"" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.118401 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e111a15d-a0c1-470b-9723-e4bbcc2265c1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.118413 4770 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/517c9323-8c40-413a-92a8-9dd833dbfc2c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.118424 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e111a15d-a0c1-470b-9723-e4bbcc2265c1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.122349 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/517c9323-8c40-413a-92a8-9dd833dbfc2c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "517c9323-8c40-413a-92a8-9dd833dbfc2c" (UID: "517c9323-8c40-413a-92a8-9dd833dbfc2c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.219452 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/517c9323-8c40-413a-92a8-9dd833dbfc2c-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.633519 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"517c9323-8c40-413a-92a8-9dd833dbfc2c","Type":"ContainerDied","Data":"5dad8233878c85ad282f4541b0c29daeabe08626f2672bda566b907769415240"} Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.633595 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dad8233878c85ad282f4541b0c29daeabe08626f2672bda566b907769415240" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.633677 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.638818 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" event={"ID":"e111a15d-a0c1-470b-9723-e4bbcc2265c1","Type":"ContainerDied","Data":"1dc6b69f1c77f89981e16f02ac2ac741af3f597b1584ac24bad8a9f4eeca9384"} Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.638891 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dc6b69f1c77f89981e16f02ac2ac741af3f597b1584ac24bad8a9f4eeca9384" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.638981 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph" Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.663371 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4e54166e-b9e8-4136-9f06-09e8480555b7","Type":"ContainerStarted","Data":"4fc4123bbb5aaa54b13200f790682482a5c654979ab6b5666b09deed55b92f9e"} Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.811403 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:45 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:45 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:45 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:45 crc kubenswrapper[4770]: I1004 03:05:45.811509 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:46 crc kubenswrapper[4770]: I1004 03:05:46.683263 4770 generic.go:334] "Generic (PLEG): container finished" podID="4e54166e-b9e8-4136-9f06-09e8480555b7" containerID="4fc4123bbb5aaa54b13200f790682482a5c654979ab6b5666b09deed55b92f9e" exitCode=0 Oct 04 03:05:46 crc kubenswrapper[4770]: I1004 03:05:46.683839 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4e54166e-b9e8-4136-9f06-09e8480555b7","Type":"ContainerDied","Data":"4fc4123bbb5aaa54b13200f790682482a5c654979ab6b5666b09deed55b92f9e"} Oct 04 03:05:46 crc kubenswrapper[4770]: I1004 03:05:46.803212 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:46 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:46 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:46 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:46 crc kubenswrapper[4770]: I1004 03:05:46.803284 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:47 crc kubenswrapper[4770]: I1004 03:05:47.802972 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:47 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:47 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:47 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:47 crc kubenswrapper[4770]: I1004 03:05:47.803141 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:47 crc kubenswrapper[4770]: I1004 03:05:47.850091 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-x2rjg" Oct 04 03:05:48 crc kubenswrapper[4770]: I1004 03:05:48.804332 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:48 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:48 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:48 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:48 crc kubenswrapper[4770]: I1004 03:05:48.804414 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:49 crc kubenswrapper[4770]: I1004 03:05:49.802512 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:49 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:49 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:49 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:49 crc kubenswrapper[4770]: I1004 03:05:49.803058 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:50 crc kubenswrapper[4770]: I1004 03:05:50.803407 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:50 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:50 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:50 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:50 crc kubenswrapper[4770]: I1004 03:05:50.803486 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:51 crc kubenswrapper[4770]: I1004 03:05:51.804442 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:51 crc kubenswrapper[4770]: [-]has-synced failed: reason withheld Oct 04 03:05:51 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:51 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:51 crc kubenswrapper[4770]: I1004 03:05:51.804956 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:52 crc kubenswrapper[4770]: I1004 03:05:52.369539 4770 patch_prober.go:28] interesting pod/console-f9d7485db-xwhz4 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 04 03:05:52 crc kubenswrapper[4770]: I1004 03:05:52.369598 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-xwhz4" podUID="2b45833c-2e11-428e-8131-1ea4676a794f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 04 03:05:52 crc kubenswrapper[4770]: I1004 03:05:52.375935 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-lwhtk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 03:05:52 crc kubenswrapper[4770]: I1004 03:05:52.375966 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lwhtk" podUID="4bf7a322-09ed-493a-bd28-2639f15402f5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 03:05:52 crc kubenswrapper[4770]: I1004 03:05:52.376213 4770 patch_prober.go:28] interesting pod/downloads-7954f5f757-lwhtk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 03:05:52 crc kubenswrapper[4770]: I1004 03:05:52.376236 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lwhtk" podUID="4bf7a322-09ed-493a-bd28-2639f15402f5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 03:05:52 crc kubenswrapper[4770]: I1004 03:05:52.804607 4770 patch_prober.go:28] interesting pod/router-default-5444994796-ggg8j container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:05:52 crc kubenswrapper[4770]: [+]has-synced ok Oct 04 03:05:52 crc kubenswrapper[4770]: [+]process-running ok Oct 04 03:05:52 crc kubenswrapper[4770]: healthz check failed Oct 04 03:05:52 crc kubenswrapper[4770]: I1004 03:05:52.804711 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ggg8j" podUID="03505445-42fb-4015-97d4-6d41167cadd1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:05:53 crc kubenswrapper[4770]: I1004 03:05:53.803006 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:53 crc kubenswrapper[4770]: I1004 03:05:53.806710 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ggg8j" Oct 04 03:05:53 crc kubenswrapper[4770]: I1004 03:05:53.921419 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:53 crc kubenswrapper[4770]: I1004 03:05:53.969254 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7fe8aa-4b62-47da-8aa7-0d31dcdba457-metrics-certs\") pod \"network-metrics-daemon-krp9c\" (UID: \"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457\") " pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.013755 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krp9c" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.110056 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.231206 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e54166e-b9e8-4136-9f06-09e8480555b7-kube-api-access\") pod \"4e54166e-b9e8-4136-9f06-09e8480555b7\" (UID: \"4e54166e-b9e8-4136-9f06-09e8480555b7\") " Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.231464 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e54166e-b9e8-4136-9f06-09e8480555b7-kubelet-dir\") pod \"4e54166e-b9e8-4136-9f06-09e8480555b7\" (UID: \"4e54166e-b9e8-4136-9f06-09e8480555b7\") " Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.231883 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e54166e-b9e8-4136-9f06-09e8480555b7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4e54166e-b9e8-4136-9f06-09e8480555b7" (UID: "4e54166e-b9e8-4136-9f06-09e8480555b7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.237864 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e54166e-b9e8-4136-9f06-09e8480555b7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4e54166e-b9e8-4136-9f06-09e8480555b7" (UID: "4e54166e-b9e8-4136-9f06-09e8480555b7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.333643 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e54166e-b9e8-4136-9f06-09e8480555b7-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.333685 4770 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e54166e-b9e8-4136-9f06-09e8480555b7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.771380 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4e54166e-b9e8-4136-9f06-09e8480555b7","Type":"ContainerDied","Data":"dbe5363c01869ff218ebf97d58173e70524043ace496a13b7f5f5e634be228ea"} Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.771424 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbe5363c01869ff218ebf97d58173e70524043ace496a13b7f5f5e634be228ea" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.771435 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.778076 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-n56bx_3b6ec72f-1166-4e03-815a-75d256d9ab5a/cluster-samples-operator/0.log" Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.778143 4770 generic.go:334] "Generic (PLEG): container finished" podID="3b6ec72f-1166-4e03-815a-75d256d9ab5a" containerID="3a6a9ecefb402c1810b76a8e7c8ab5e4eb159d032ec17225de85b0d7fa577874" exitCode=2 Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.778228 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" event={"ID":"3b6ec72f-1166-4e03-815a-75d256d9ab5a","Type":"ContainerDied","Data":"3a6a9ecefb402c1810b76a8e7c8ab5e4eb159d032ec17225de85b0d7fa577874"} Oct 04 03:05:54 crc kubenswrapper[4770]: I1004 03:05:54.779562 4770 scope.go:117] "RemoveContainer" containerID="3a6a9ecefb402c1810b76a8e7c8ab5e4eb159d032ec17225de85b0d7fa577874" Oct 04 03:06:01 crc kubenswrapper[4770]: I1004 03:06:01.273141 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:06:01 crc kubenswrapper[4770]: I1004 03:06:01.795492 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:06:01 crc kubenswrapper[4770]: I1004 03:06:01.795566 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:06:02 crc kubenswrapper[4770]: I1004 03:06:02.385046 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-lwhtk" Oct 04 03:06:02 crc kubenswrapper[4770]: I1004 03:06:02.600511 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:06:02 crc kubenswrapper[4770]: I1004 03:06:02.604787 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:06:12 crc kubenswrapper[4770]: I1004 03:06:12.876511 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-g8rqs" Oct 04 03:06:13 crc kubenswrapper[4770]: E1004 03:06:13.536439 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 03:06:13 crc kubenswrapper[4770]: E1004 03:06:13.536698 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l4fdc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-dcs7t_openshift-marketplace(30e80d37-aeb7-4446-852c-3a3ac2b78bb8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:06:13 crc kubenswrapper[4770]: E1004 03:06:13.537883 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-dcs7t" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" Oct 04 03:06:14 crc kubenswrapper[4770]: E1004 03:06:14.863036 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-dcs7t" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" Oct 04 03:06:14 crc kubenswrapper[4770]: E1004 03:06:14.940127 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 03:06:14 crc kubenswrapper[4770]: E1004 03:06:14.940289 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wtfg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-97vj5_openshift-marketplace(4c3242ed-a5f8-4078-b2fc-1031ce5e92b1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:06:14 crc kubenswrapper[4770]: E1004 03:06:14.941571 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-97vj5" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" Oct 04 03:06:14 crc kubenswrapper[4770]: E1004 03:06:14.945389 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 03:06:14 crc kubenswrapper[4770]: E1004 03:06:14.945580 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s28jc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-m4jht_openshift-marketplace(d6d87160-c0c6-4cb0-8561-13988e9c5395): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:06:14 crc kubenswrapper[4770]: E1004 03:06:14.946989 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-m4jht" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" Oct 04 03:06:18 crc kubenswrapper[4770]: E1004 03:06:18.031713 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-97vj5" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" Oct 04 03:06:18 crc kubenswrapper[4770]: E1004 03:06:18.031760 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-m4jht" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" Oct 04 03:06:18 crc kubenswrapper[4770]: E1004 03:06:18.116287 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 03:06:18 crc kubenswrapper[4770]: E1004 03:06:18.116538 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7jzwc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pvjjj_openshift-marketplace(f8207392-9683-4278-a7d5-8e6af62fbf61): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:06:18 crc kubenswrapper[4770]: E1004 03:06:18.118077 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pvjjj" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" Oct 04 03:06:18 crc kubenswrapper[4770]: E1004 03:06:18.162328 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 03:06:18 crc kubenswrapper[4770]: E1004 03:06:18.163094 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w7skx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zzdjv_openshift-marketplace(ac60a2ae-8790-4e3d-a2c0-ba394a8140a6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:06:18 crc kubenswrapper[4770]: E1004 03:06:18.164494 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zzdjv" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" Oct 04 03:06:18 crc kubenswrapper[4770]: I1004 03:06:18.601553 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:06:20 crc kubenswrapper[4770]: E1004 03:06:20.666443 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zzdjv" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" Oct 04 03:06:20 crc kubenswrapper[4770]: E1004 03:06:20.666504 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-pvjjj" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" Oct 04 03:06:20 crc kubenswrapper[4770]: E1004 03:06:20.760453 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 03:06:20 crc kubenswrapper[4770]: E1004 03:06:20.760789 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9v5g9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rnkr7_openshift-marketplace(a9dbdce0-d5c5-4b60-9d4b-745beca12ee8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:06:20 crc kubenswrapper[4770]: E1004 03:06:20.764907 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rnkr7" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" Oct 04 03:06:21 crc kubenswrapper[4770]: I1004 03:06:21.098765 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-krp9c"] Oct 04 03:06:21 crc kubenswrapper[4770]: E1004 03:06:21.623719 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rnkr7" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" Oct 04 03:06:21 crc kubenswrapper[4770]: W1004 03:06:21.638291 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e7fe8aa_4b62_47da_8aa7_0d31dcdba457.slice/crio-8d77ac1292c08cc458fabc0afebfb1b906a74d1cbc52487ffb58a2d9dab44652 WatchSource:0}: Error finding container 8d77ac1292c08cc458fabc0afebfb1b906a74d1cbc52487ffb58a2d9dab44652: Status 404 returned error can't find the container with id 8d77ac1292c08cc458fabc0afebfb1b906a74d1cbc52487ffb58a2d9dab44652 Oct 04 03:06:21 crc kubenswrapper[4770]: E1004 03:06:21.692543 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 03:06:21 crc kubenswrapper[4770]: E1004 03:06:21.692823 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4pmd4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bk4rd_openshift-marketplace(400cc27a-45ec-4164-9af6-7c120fff6351): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:06:21 crc kubenswrapper[4770]: E1004 03:06:21.694104 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bk4rd" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" Oct 04 03:06:21 crc kubenswrapper[4770]: E1004 03:06:21.726972 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 03:06:21 crc kubenswrapper[4770]: E1004 03:06:21.727196 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hp6xb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-n9zkn_openshift-marketplace(bba2dc2c-e45e-43c2-8d02-ee01bf41143a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:06:21 crc kubenswrapper[4770]: E1004 03:06:21.728371 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-n9zkn" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" Oct 04 03:06:21 crc kubenswrapper[4770]: I1004 03:06:21.960108 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-n56bx_3b6ec72f-1166-4e03-815a-75d256d9ab5a/cluster-samples-operator/0.log" Oct 04 03:06:21 crc kubenswrapper[4770]: I1004 03:06:21.960448 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n56bx" event={"ID":"3b6ec72f-1166-4e03-815a-75d256d9ab5a","Type":"ContainerStarted","Data":"426f6455bf51a2ea0e0c77c82694a730e78a640845771c2a1f701459edccf65e"} Oct 04 03:06:21 crc kubenswrapper[4770]: I1004 03:06:21.963213 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krp9c" event={"ID":"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457","Type":"ContainerStarted","Data":"aade9a9abcc7d0f70a2e19bcc493ce7e42a98884a177ab7b6678d839142a36eb"} Oct 04 03:06:21 crc kubenswrapper[4770]: I1004 03:06:21.963234 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krp9c" event={"ID":"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457","Type":"ContainerStarted","Data":"8d77ac1292c08cc458fabc0afebfb1b906a74d1cbc52487ffb58a2d9dab44652"} Oct 04 03:06:21 crc kubenswrapper[4770]: E1004 03:06:21.964270 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bk4rd" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" Oct 04 03:06:21 crc kubenswrapper[4770]: E1004 03:06:21.965654 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-n9zkn" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" Oct 04 03:06:22 crc kubenswrapper[4770]: I1004 03:06:22.972420 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krp9c" event={"ID":"5e7fe8aa-4b62-47da-8aa7-0d31dcdba457","Type":"ContainerStarted","Data":"d4f559f9bdde8b147b99a64ca99415d7d468d0607c5a7725ed384c34c808c7d6"} Oct 04 03:06:22 crc kubenswrapper[4770]: I1004 03:06:22.991100 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-krp9c" podStartSLOduration=172.991070349 podStartE2EDuration="2m52.991070349s" podCreationTimestamp="2025-10-04 03:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:06:22.987912671 +0000 UTC m=+194.279922403" watchObservedRunningTime="2025-10-04 03:06:22.991070349 +0000 UTC m=+194.283080101" Oct 04 03:06:31 crc kubenswrapper[4770]: I1004 03:06:31.795837 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:06:31 crc kubenswrapper[4770]: I1004 03:06:31.796740 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:06:38 crc kubenswrapper[4770]: I1004 03:06:38.083123 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4jht" event={"ID":"d6d87160-c0c6-4cb0-8561-13988e9c5395","Type":"ContainerStarted","Data":"49a6c4b91ce5a1f4247a4216fe641c4f1aef8a4eb6025ddd89e98086b0c6446f"} Oct 04 03:06:38 crc kubenswrapper[4770]: I1004 03:06:38.085264 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dcs7t" event={"ID":"30e80d37-aeb7-4446-852c-3a3ac2b78bb8","Type":"ContainerStarted","Data":"0afc29f7d9fc01fdee1497b4709f6133045f26ea902e4d063a0369c48053301a"} Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.097095 4770 generic.go:334] "Generic (PLEG): container finished" podID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerID="49a6c4b91ce5a1f4247a4216fe641c4f1aef8a4eb6025ddd89e98086b0c6446f" exitCode=0 Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.097196 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4jht" event={"ID":"d6d87160-c0c6-4cb0-8561-13988e9c5395","Type":"ContainerDied","Data":"49a6c4b91ce5a1f4247a4216fe641c4f1aef8a4eb6025ddd89e98086b0c6446f"} Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.104792 4770 generic.go:334] "Generic (PLEG): container finished" podID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerID="d4ad73927db7a50979a4e55fb419a5d125d33dfe98043726424d5fc565ff2c07" exitCode=0 Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.104924 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9zkn" event={"ID":"bba2dc2c-e45e-43c2-8d02-ee01bf41143a","Type":"ContainerDied","Data":"d4ad73927db7a50979a4e55fb419a5d125d33dfe98043726424d5fc565ff2c07"} Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.120828 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97vj5" event={"ID":"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1","Type":"ContainerStarted","Data":"5f98319d83f952f87a95a7af07cdabf75782253dc8d438b6a17823eca8193974"} Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.122476 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzdjv" event={"ID":"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6","Type":"ContainerStarted","Data":"f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99"} Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.133069 4770 generic.go:334] "Generic (PLEG): container finished" podID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerID="0afc29f7d9fc01fdee1497b4709f6133045f26ea902e4d063a0369c48053301a" exitCode=0 Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.133178 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dcs7t" event={"ID":"30e80d37-aeb7-4446-852c-3a3ac2b78bb8","Type":"ContainerDied","Data":"0afc29f7d9fc01fdee1497b4709f6133045f26ea902e4d063a0369c48053301a"} Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.152852 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk4rd" event={"ID":"400cc27a-45ec-4164-9af6-7c120fff6351","Type":"ContainerStarted","Data":"1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb"} Oct 04 03:06:39 crc kubenswrapper[4770]: I1004 03:06:39.244673 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dcs7t" podStartSLOduration=3.812850765 podStartE2EDuration="1m1.24464471s" podCreationTimestamp="2025-10-04 03:05:38 +0000 UTC" firstStartedPulling="2025-10-04 03:05:41.349416442 +0000 UTC m=+152.641426154" lastFinishedPulling="2025-10-04 03:06:38.781210387 +0000 UTC m=+210.073220099" observedRunningTime="2025-10-04 03:06:39.241467865 +0000 UTC m=+210.533477577" watchObservedRunningTime="2025-10-04 03:06:39.24464471 +0000 UTC m=+210.536654422" Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.185226 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4jht" event={"ID":"d6d87160-c0c6-4cb0-8561-13988e9c5395","Type":"ContainerStarted","Data":"0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62"} Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.191785 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9zkn" event={"ID":"bba2dc2c-e45e-43c2-8d02-ee01bf41143a","Type":"ContainerStarted","Data":"741c73b49d6e423d55caca41271806e8ccd46dfb5ac23bf8431dadf17e3f6973"} Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.194021 4770 generic.go:334] "Generic (PLEG): container finished" podID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerID="5f98319d83f952f87a95a7af07cdabf75782253dc8d438b6a17823eca8193974" exitCode=0 Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.194064 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97vj5" event={"ID":"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1","Type":"ContainerDied","Data":"5f98319d83f952f87a95a7af07cdabf75782253dc8d438b6a17823eca8193974"} Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.196990 4770 generic.go:334] "Generic (PLEG): container finished" podID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerID="f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99" exitCode=0 Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.197071 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzdjv" event={"ID":"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6","Type":"ContainerDied","Data":"f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99"} Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.201441 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dcs7t" event={"ID":"30e80d37-aeb7-4446-852c-3a3ac2b78bb8","Type":"ContainerStarted","Data":"b73c0d574135fb618dec4515ea9891db47ad4a2e00d7ccbcb15640ede0599544"} Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.207119 4770 generic.go:334] "Generic (PLEG): container finished" podID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerID="8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95" exitCode=0 Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.207221 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rnkr7" event={"ID":"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8","Type":"ContainerDied","Data":"8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95"} Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.209923 4770 generic.go:334] "Generic (PLEG): container finished" podID="400cc27a-45ec-4164-9af6-7c120fff6351" containerID="1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb" exitCode=0 Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.209996 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk4rd" event={"ID":"400cc27a-45ec-4164-9af6-7c120fff6351","Type":"ContainerDied","Data":"1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb"} Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.219825 4770 generic.go:334] "Generic (PLEG): container finished" podID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerID="e3d5d35839a4aaad2bd9775847734e6c7ff9cb9bdc3c4e45f74603e0f99a79d4" exitCode=0 Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.219884 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjjj" event={"ID":"f8207392-9683-4278-a7d5-8e6af62fbf61","Type":"ContainerDied","Data":"e3d5d35839a4aaad2bd9775847734e6c7ff9cb9bdc3c4e45f74603e0f99a79d4"} Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.226785 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m4jht" podStartSLOduration=3.179698695 podStartE2EDuration="1m1.22675326s" podCreationTimestamp="2025-10-04 03:05:39 +0000 UTC" firstStartedPulling="2025-10-04 03:05:41.487872908 +0000 UTC m=+152.779882620" lastFinishedPulling="2025-10-04 03:06:39.534927453 +0000 UTC m=+210.826937185" observedRunningTime="2025-10-04 03:06:40.211394053 +0000 UTC m=+211.503403765" watchObservedRunningTime="2025-10-04 03:06:40.22675326 +0000 UTC m=+211.518762972" Oct 04 03:06:40 crc kubenswrapper[4770]: I1004 03:06:40.316190 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n9zkn" podStartSLOduration=3.030880129 podStartE2EDuration="1m0.31616041s" podCreationTimestamp="2025-10-04 03:05:40 +0000 UTC" firstStartedPulling="2025-10-04 03:05:42.486658874 +0000 UTC m=+153.778668586" lastFinishedPulling="2025-10-04 03:06:39.771939155 +0000 UTC m=+211.063948867" observedRunningTime="2025-10-04 03:06:40.311306871 +0000 UTC m=+211.603316593" watchObservedRunningTime="2025-10-04 03:06:40.31616041 +0000 UTC m=+211.608170152" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.039659 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.039958 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.226909 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjjj" event={"ID":"f8207392-9683-4278-a7d5-8e6af62fbf61","Type":"ContainerStarted","Data":"cd332699ab2404ce4b67af9670fd442e8c91c75bee97f6455a41444f46ac1361"} Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.228859 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97vj5" event={"ID":"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1","Type":"ContainerStarted","Data":"694a1c93e474146772291544f0f266ff5f2c6f90859b9609eb328738deb7dd47"} Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.231298 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzdjv" event={"ID":"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6","Type":"ContainerStarted","Data":"a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f"} Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.232741 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rnkr7" event={"ID":"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8","Type":"ContainerStarted","Data":"55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8"} Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.235637 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk4rd" event={"ID":"400cc27a-45ec-4164-9af6-7c120fff6351","Type":"ContainerStarted","Data":"42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87"} Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.259151 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pvjjj" podStartSLOduration=2.2352181939999998 podStartE2EDuration="59.259127603s" podCreationTimestamp="2025-10-04 03:05:42 +0000 UTC" firstStartedPulling="2025-10-04 03:05:43.585263297 +0000 UTC m=+154.877273009" lastFinishedPulling="2025-10-04 03:06:40.609172686 +0000 UTC m=+211.901182418" observedRunningTime="2025-10-04 03:06:41.258106346 +0000 UTC m=+212.550116068" watchObservedRunningTime="2025-10-04 03:06:41.259127603 +0000 UTC m=+212.551137325" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.261399 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.308634 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rnkr7" podStartSLOduration=2.779813339 podStartE2EDuration="1m2.308611515s" podCreationTimestamp="2025-10-04 03:05:39 +0000 UTC" firstStartedPulling="2025-10-04 03:05:41.361736457 +0000 UTC m=+152.653746169" lastFinishedPulling="2025-10-04 03:06:40.890534633 +0000 UTC m=+212.182544345" observedRunningTime="2025-10-04 03:06:41.30618616 +0000 UTC m=+212.598195882" watchObservedRunningTime="2025-10-04 03:06:41.308611515 +0000 UTC m=+212.600621227" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.310060 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-97vj5" podStartSLOduration=4.129587737 podStartE2EDuration="1m3.310052753s" podCreationTimestamp="2025-10-04 03:05:38 +0000 UTC" firstStartedPulling="2025-10-04 03:05:41.487788946 +0000 UTC m=+152.779798658" lastFinishedPulling="2025-10-04 03:06:40.668253952 +0000 UTC m=+211.960263674" observedRunningTime="2025-10-04 03:06:41.287470325 +0000 UTC m=+212.579480037" watchObservedRunningTime="2025-10-04 03:06:41.310052753 +0000 UTC m=+212.602062465" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.356737 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bk4rd" podStartSLOduration=2.139587385 podStartE2EDuration="1m0.35671787s" podCreationTimestamp="2025-10-04 03:05:41 +0000 UTC" firstStartedPulling="2025-10-04 03:05:42.491257641 +0000 UTC m=+153.783267353" lastFinishedPulling="2025-10-04 03:06:40.708388126 +0000 UTC m=+212.000397838" observedRunningTime="2025-10-04 03:06:41.35521024 +0000 UTC m=+212.647219952" watchObservedRunningTime="2025-10-04 03:06:41.35671787 +0000 UTC m=+212.648727582" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.360121 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zzdjv" podStartSLOduration=3.165286852 podStartE2EDuration="59.36010785s" podCreationTimestamp="2025-10-04 03:05:42 +0000 UTC" firstStartedPulling="2025-10-04 03:05:44.599982529 +0000 UTC m=+155.891992241" lastFinishedPulling="2025-10-04 03:06:40.794803527 +0000 UTC m=+212.086813239" observedRunningTime="2025-10-04 03:06:41.329314314 +0000 UTC m=+212.621324046" watchObservedRunningTime="2025-10-04 03:06:41.36010785 +0000 UTC m=+212.652117562" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.395253 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:06:41 crc kubenswrapper[4770]: I1004 03:06:41.395314 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:06:42 crc kubenswrapper[4770]: I1004 03:06:42.392236 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:06:42 crc kubenswrapper[4770]: I1004 03:06:42.392706 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:06:42 crc kubenswrapper[4770]: I1004 03:06:42.445661 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-bk4rd" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" containerName="registry-server" probeResult="failure" output=< Oct 04 03:06:42 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 03:06:42 crc kubenswrapper[4770]: > Oct 04 03:06:42 crc kubenswrapper[4770]: I1004 03:06:42.836602 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:06:42 crc kubenswrapper[4770]: I1004 03:06:42.837192 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:06:43 crc kubenswrapper[4770]: I1004 03:06:43.459742 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pvjjj" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerName="registry-server" probeResult="failure" output=< Oct 04 03:06:43 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 03:06:43 crc kubenswrapper[4770]: > Oct 04 03:06:43 crc kubenswrapper[4770]: I1004 03:06:43.880140 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zzdjv" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerName="registry-server" probeResult="failure" output=< Oct 04 03:06:43 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 03:06:43 crc kubenswrapper[4770]: > Oct 04 03:06:44 crc kubenswrapper[4770]: I1004 03:06:44.286581 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r8rk"] Oct 04 03:06:48 crc kubenswrapper[4770]: I1004 03:06:48.885478 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:06:48 crc kubenswrapper[4770]: I1004 03:06:48.886050 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:06:48 crc kubenswrapper[4770]: I1004 03:06:48.947595 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.201129 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.201905 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.244999 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.349931 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.350769 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.510863 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.510943 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.561094 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.690588 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.690652 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:06:49 crc kubenswrapper[4770]: I1004 03:06:49.740860 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:06:50 crc kubenswrapper[4770]: I1004 03:06:50.343339 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:06:50 crc kubenswrapper[4770]: I1004 03:06:50.353762 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:06:50 crc kubenswrapper[4770]: I1004 03:06:50.581665 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-97vj5"] Oct 04 03:06:51 crc kubenswrapper[4770]: I1004 03:06:51.098491 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:06:51 crc kubenswrapper[4770]: I1004 03:06:51.309743 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-97vj5" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerName="registry-server" containerID="cri-o://694a1c93e474146772291544f0f266ff5f2c6f90859b9609eb328738deb7dd47" gracePeriod=2 Oct 04 03:06:51 crc kubenswrapper[4770]: I1004 03:06:51.486798 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:06:51 crc kubenswrapper[4770]: I1004 03:06:51.530835 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:06:51 crc kubenswrapper[4770]: I1004 03:06:51.982601 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rnkr7"] Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.333398 4770 generic.go:334] "Generic (PLEG): container finished" podID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerID="694a1c93e474146772291544f0f266ff5f2c6f90859b9609eb328738deb7dd47" exitCode=0 Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.333430 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97vj5" event={"ID":"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1","Type":"ContainerDied","Data":"694a1c93e474146772291544f0f266ff5f2c6f90859b9609eb328738deb7dd47"} Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.333710 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rnkr7" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerName="registry-server" containerID="cri-o://55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8" gracePeriod=2 Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.441635 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.493371 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.550136 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.617186 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtfg6\" (UniqueName: \"kubernetes.io/projected/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-kube-api-access-wtfg6\") pod \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.617280 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-catalog-content\") pod \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.617351 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-utilities\") pod \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\" (UID: \"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1\") " Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.619090 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-utilities" (OuterVolumeSpecName: "utilities") pod "4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" (UID: "4c3242ed-a5f8-4078-b2fc-1031ce5e92b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.626463 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-kube-api-access-wtfg6" (OuterVolumeSpecName: "kube-api-access-wtfg6") pod "4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" (UID: "4c3242ed-a5f8-4078-b2fc-1031ce5e92b1"). InnerVolumeSpecName "kube-api-access-wtfg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.669090 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" (UID: "4c3242ed-a5f8-4078-b2fc-1031ce5e92b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.719316 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtfg6\" (UniqueName: \"kubernetes.io/projected/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-kube-api-access-wtfg6\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.719347 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.719362 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.878019 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:06:52 crc kubenswrapper[4770]: I1004 03:06:52.925490 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.288551 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.327187 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-utilities\") pod \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.327307 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v5g9\" (UniqueName: \"kubernetes.io/projected/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-kube-api-access-9v5g9\") pod \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.327345 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-catalog-content\") pod \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\" (UID: \"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8\") " Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.328124 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-utilities" (OuterVolumeSpecName: "utilities") pod "a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" (UID: "a9dbdce0-d5c5-4b60-9d4b-745beca12ee8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.330654 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-kube-api-access-9v5g9" (OuterVolumeSpecName: "kube-api-access-9v5g9") pod "a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" (UID: "a9dbdce0-d5c5-4b60-9d4b-745beca12ee8"). InnerVolumeSpecName "kube-api-access-9v5g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.344420 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-97vj5" event={"ID":"4c3242ed-a5f8-4078-b2fc-1031ce5e92b1","Type":"ContainerDied","Data":"f7b8e2ceaa4dec5684455baed16bc1a1cd055d5bad952c94b7d2f0b653b1e59b"} Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.344562 4770 scope.go:117] "RemoveContainer" containerID="694a1c93e474146772291544f0f266ff5f2c6f90859b9609eb328738deb7dd47" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.344567 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-97vj5" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.348210 4770 generic.go:334] "Generic (PLEG): container finished" podID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerID="55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8" exitCode=0 Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.348274 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rnkr7" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.348337 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rnkr7" event={"ID":"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8","Type":"ContainerDied","Data":"55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8"} Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.348375 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rnkr7" event={"ID":"a9dbdce0-d5c5-4b60-9d4b-745beca12ee8","Type":"ContainerDied","Data":"f5f3b765ec21011bd2c2374ad7383a5790fc383cda592613efd6c8844b3ab89a"} Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.361823 4770 scope.go:117] "RemoveContainer" containerID="5f98319d83f952f87a95a7af07cdabf75782253dc8d438b6a17823eca8193974" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.380069 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-97vj5"] Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.380646 4770 scope.go:117] "RemoveContainer" containerID="47f4e67aa2ca7bcf80ffe0f1a8f1cac3afe778c54efd21f5432355bb52b0146f" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.381645 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-97vj5"] Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.383172 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" (UID: "a9dbdce0-d5c5-4b60-9d4b-745beca12ee8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.420559 4770 scope.go:117] "RemoveContainer" containerID="55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.428385 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.428415 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.428427 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v5g9\" (UniqueName: \"kubernetes.io/projected/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8-kube-api-access-9v5g9\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.434930 4770 scope.go:117] "RemoveContainer" containerID="8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.452694 4770 scope.go:117] "RemoveContainer" containerID="a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.469904 4770 scope.go:117] "RemoveContainer" containerID="55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8" Oct 04 03:06:53 crc kubenswrapper[4770]: E1004 03:06:53.470351 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8\": container with ID starting with 55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8 not found: ID does not exist" containerID="55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.470389 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8"} err="failed to get container status \"55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8\": rpc error: code = NotFound desc = could not find container \"55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8\": container with ID starting with 55aece8c57f153bbd60db7c2ca0c138d8aa2c6875a2b06933cc72efe311fc4f8 not found: ID does not exist" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.470438 4770 scope.go:117] "RemoveContainer" containerID="8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95" Oct 04 03:06:53 crc kubenswrapper[4770]: E1004 03:06:53.470772 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95\": container with ID starting with 8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95 not found: ID does not exist" containerID="8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.470794 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95"} err="failed to get container status \"8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95\": rpc error: code = NotFound desc = could not find container \"8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95\": container with ID starting with 8479bbcbba196b6ded91772dc6bc9a620e379ebeaa8f03b81594fad0f3744b95 not found: ID does not exist" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.470807 4770 scope.go:117] "RemoveContainer" containerID="a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3" Oct 04 03:06:53 crc kubenswrapper[4770]: E1004 03:06:53.471301 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3\": container with ID starting with a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3 not found: ID does not exist" containerID="a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.471321 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3"} err="failed to get container status \"a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3\": rpc error: code = NotFound desc = could not find container \"a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3\": container with ID starting with a933680224525aaaa91ab4dadc2c05be38c997960c3c5f81e0880ab195136ef3 not found: ID does not exist" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.681620 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" path="/var/lib/kubelet/pods/4c3242ed-a5f8-4078-b2fc-1031ce5e92b1/volumes" Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.682216 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rnkr7"] Oct 04 03:06:53 crc kubenswrapper[4770]: I1004 03:06:53.683211 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rnkr7"] Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.379787 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk4rd"] Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.380024 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bk4rd" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" containerName="registry-server" containerID="cri-o://42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87" gracePeriod=2 Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.779018 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.850301 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pmd4\" (UniqueName: \"kubernetes.io/projected/400cc27a-45ec-4164-9af6-7c120fff6351-kube-api-access-4pmd4\") pod \"400cc27a-45ec-4164-9af6-7c120fff6351\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.850374 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-catalog-content\") pod \"400cc27a-45ec-4164-9af6-7c120fff6351\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.851327 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-utilities\") pod \"400cc27a-45ec-4164-9af6-7c120fff6351\" (UID: \"400cc27a-45ec-4164-9af6-7c120fff6351\") " Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.852075 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-utilities" (OuterVolumeSpecName: "utilities") pod "400cc27a-45ec-4164-9af6-7c120fff6351" (UID: "400cc27a-45ec-4164-9af6-7c120fff6351"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.859205 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/400cc27a-45ec-4164-9af6-7c120fff6351-kube-api-access-4pmd4" (OuterVolumeSpecName: "kube-api-access-4pmd4") pod "400cc27a-45ec-4164-9af6-7c120fff6351" (UID: "400cc27a-45ec-4164-9af6-7c120fff6351"). InnerVolumeSpecName "kube-api-access-4pmd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.869645 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "400cc27a-45ec-4164-9af6-7c120fff6351" (UID: "400cc27a-45ec-4164-9af6-7c120fff6351"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.952696 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.952737 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pmd4\" (UniqueName: \"kubernetes.io/projected/400cc27a-45ec-4164-9af6-7c120fff6351-kube-api-access-4pmd4\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:54 crc kubenswrapper[4770]: I1004 03:06:54.952750 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/400cc27a-45ec-4164-9af6-7c120fff6351-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.369222 4770 generic.go:334] "Generic (PLEG): container finished" podID="400cc27a-45ec-4164-9af6-7c120fff6351" containerID="42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87" exitCode=0 Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.369306 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk4rd" event={"ID":"400cc27a-45ec-4164-9af6-7c120fff6351","Type":"ContainerDied","Data":"42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87"} Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.369473 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk4rd" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.369495 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk4rd" event={"ID":"400cc27a-45ec-4164-9af6-7c120fff6351","Type":"ContainerDied","Data":"1ecdb793ad527a3635d4c0b97107475f94e0d1938c9e063cca2985424c4098d6"} Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.369544 4770 scope.go:117] "RemoveContainer" containerID="42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.391893 4770 scope.go:117] "RemoveContainer" containerID="1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.410357 4770 scope.go:117] "RemoveContainer" containerID="df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.410678 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk4rd"] Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.413437 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk4rd"] Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.439985 4770 scope.go:117] "RemoveContainer" containerID="42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87" Oct 04 03:06:55 crc kubenswrapper[4770]: E1004 03:06:55.440564 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87\": container with ID starting with 42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87 not found: ID does not exist" containerID="42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.440615 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87"} err="failed to get container status \"42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87\": rpc error: code = NotFound desc = could not find container \"42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87\": container with ID starting with 42c02c38122ed232f5aa3e81adca7cf2fafb21e559c03f7fe09b2d87a6a5ee87 not found: ID does not exist" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.440653 4770 scope.go:117] "RemoveContainer" containerID="1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb" Oct 04 03:06:55 crc kubenswrapper[4770]: E1004 03:06:55.441161 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb\": container with ID starting with 1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb not found: ID does not exist" containerID="1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.441183 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb"} err="failed to get container status \"1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb\": rpc error: code = NotFound desc = could not find container \"1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb\": container with ID starting with 1b7688052acefb8ca2b5c608d962fc9f38a162e54bee4a92e7d8696cadbdddeb not found: ID does not exist" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.441198 4770 scope.go:117] "RemoveContainer" containerID="df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0" Oct 04 03:06:55 crc kubenswrapper[4770]: E1004 03:06:55.441508 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0\": container with ID starting with df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0 not found: ID does not exist" containerID="df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.441534 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0"} err="failed to get container status \"df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0\": rpc error: code = NotFound desc = could not find container \"df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0\": container with ID starting with df69f75f49a7fdde93744a90b9474bdd0d004ca7755fb856e38cc6045beb7ef0 not found: ID does not exist" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.682760 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" path="/var/lib/kubelet/pods/400cc27a-45ec-4164-9af6-7c120fff6351/volumes" Oct 04 03:06:55 crc kubenswrapper[4770]: I1004 03:06:55.683824 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" path="/var/lib/kubelet/pods/a9dbdce0-d5c5-4b60-9d4b-745beca12ee8/volumes" Oct 04 03:06:56 crc kubenswrapper[4770]: I1004 03:06:56.781920 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzdjv"] Oct 04 03:06:56 crc kubenswrapper[4770]: I1004 03:06:56.782302 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zzdjv" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerName="registry-server" containerID="cri-o://a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f" gracePeriod=2 Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.134439 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.189228 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7skx\" (UniqueName: \"kubernetes.io/projected/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-kube-api-access-w7skx\") pod \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.189307 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-utilities\") pod \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.189346 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-catalog-content\") pod \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\" (UID: \"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6\") " Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.190331 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-utilities" (OuterVolumeSpecName: "utilities") pod "ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" (UID: "ac60a2ae-8790-4e3d-a2c0-ba394a8140a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.194427 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-kube-api-access-w7skx" (OuterVolumeSpecName: "kube-api-access-w7skx") pod "ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" (UID: "ac60a2ae-8790-4e3d-a2c0-ba394a8140a6"). InnerVolumeSpecName "kube-api-access-w7skx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.283998 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" (UID: "ac60a2ae-8790-4e3d-a2c0-ba394a8140a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.291176 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7skx\" (UniqueName: \"kubernetes.io/projected/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-kube-api-access-w7skx\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.291202 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.291215 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.390580 4770 generic.go:334] "Generic (PLEG): container finished" podID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerID="a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f" exitCode=0 Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.390651 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzdjv" event={"ID":"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6","Type":"ContainerDied","Data":"a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f"} Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.390698 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzdjv" event={"ID":"ac60a2ae-8790-4e3d-a2c0-ba394a8140a6","Type":"ContainerDied","Data":"2f95f0890cb1e2708fb936668e408ea77360245e1a2e2fa43c7a4dde8dc22927"} Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.390728 4770 scope.go:117] "RemoveContainer" containerID="a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.390908 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzdjv" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.417206 4770 scope.go:117] "RemoveContainer" containerID="f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.434728 4770 scope.go:117] "RemoveContainer" containerID="cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.473990 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzdjv"] Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.475306 4770 scope.go:117] "RemoveContainer" containerID="a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f" Oct 04 03:06:57 crc kubenswrapper[4770]: E1004 03:06:57.476149 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f\": container with ID starting with a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f not found: ID does not exist" containerID="a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.476201 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f"} err="failed to get container status \"a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f\": rpc error: code = NotFound desc = could not find container \"a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f\": container with ID starting with a60b7a38df400bb5a7f39e3b08cfdf35f5eb325592a99e8ea1aacff1099e7f9f not found: ID does not exist" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.476243 4770 scope.go:117] "RemoveContainer" containerID="f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.476467 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zzdjv"] Oct 04 03:06:57 crc kubenswrapper[4770]: E1004 03:06:57.476681 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99\": container with ID starting with f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99 not found: ID does not exist" containerID="f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.476728 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99"} err="failed to get container status \"f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99\": rpc error: code = NotFound desc = could not find container \"f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99\": container with ID starting with f2bc8bce948406ee500b2e5dec4f11c26ca92b3d9abe7c6d1d176bda41d66a99 not found: ID does not exist" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.476758 4770 scope.go:117] "RemoveContainer" containerID="cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2" Oct 04 03:06:57 crc kubenswrapper[4770]: E1004 03:06:57.477211 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2\": container with ID starting with cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2 not found: ID does not exist" containerID="cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.477252 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2"} err="failed to get container status \"cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2\": rpc error: code = NotFound desc = could not find container \"cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2\": container with ID starting with cbdb5200d79c406a5ef1ab4216e51eed6438120cba49b594503054e34b925ee2 not found: ID does not exist" Oct 04 03:06:57 crc kubenswrapper[4770]: I1004 03:06:57.685155 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" path="/var/lib/kubelet/pods/ac60a2ae-8790-4e3d-a2c0-ba394a8140a6/volumes" Oct 04 03:07:01 crc kubenswrapper[4770]: I1004 03:07:01.795872 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:07:01 crc kubenswrapper[4770]: I1004 03:07:01.796055 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:07:01 crc kubenswrapper[4770]: I1004 03:07:01.796223 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:07:01 crc kubenswrapper[4770]: I1004 03:07:01.798797 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:07:01 crc kubenswrapper[4770]: I1004 03:07:01.799328 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6" gracePeriod=600 Oct 04 03:07:02 crc kubenswrapper[4770]: I1004 03:07:02.424519 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6" exitCode=0 Oct 04 03:07:02 crc kubenswrapper[4770]: I1004 03:07:02.424711 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6"} Oct 04 03:07:02 crc kubenswrapper[4770]: I1004 03:07:02.424886 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"52bc3793e50aecbadbef1d828c974de853af6a75a471a3fa3860ec13be9d7c91"} Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.320751 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" podUID="45c9685e-2941-4778-97d2-f745955fa577" containerName="oauth-openshift" containerID="cri-o://dd0d47a0ef811a3f1a95564d6f288e0a3dff34b5e663f02dadeeb9aa0a38d763" gracePeriod=15 Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.476163 4770 generic.go:334] "Generic (PLEG): container finished" podID="45c9685e-2941-4778-97d2-f745955fa577" containerID="dd0d47a0ef811a3f1a95564d6f288e0a3dff34b5e663f02dadeeb9aa0a38d763" exitCode=0 Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.476287 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" event={"ID":"45c9685e-2941-4778-97d2-f745955fa577","Type":"ContainerDied","Data":"dd0d47a0ef811a3f1a95564d6f288e0a3dff34b5e663f02dadeeb9aa0a38d763"} Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.709207 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744022 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7f54ff7574-k6bww"] Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744345 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerName="extract-content" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744360 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerName="extract-content" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744374 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerName="extract-content" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744379 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerName="extract-content" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744387 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" containerName="extract-content" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744393 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" containerName="extract-content" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744403 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerName="extract-utilities" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744409 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerName="extract-utilities" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744418 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744424 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744434 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744439 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744447 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" containerName="extract-utilities" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744454 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" containerName="extract-utilities" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744463 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerName="extract-content" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744469 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerName="extract-content" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744476 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c9685e-2941-4778-97d2-f745955fa577" containerName="oauth-openshift" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744482 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c9685e-2941-4778-97d2-f745955fa577" containerName="oauth-openshift" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744488 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerName="extract-utilities" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744494 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerName="extract-utilities" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744503 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e111a15d-a0c1-470b-9723-e4bbcc2265c1" containerName="collect-profiles" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744508 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e111a15d-a0c1-470b-9723-e4bbcc2265c1" containerName="collect-profiles" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744516 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e54166e-b9e8-4136-9f06-09e8480555b7" containerName="pruner" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744522 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e54166e-b9e8-4136-9f06-09e8480555b7" containerName="pruner" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744531 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerName="extract-utilities" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744539 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerName="extract-utilities" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744547 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517c9323-8c40-413a-92a8-9dd833dbfc2c" containerName="pruner" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744553 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="517c9323-8c40-413a-92a8-9dd833dbfc2c" containerName="pruner" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744561 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744566 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: E1004 03:07:09.744572 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744577 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744659 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="517c9323-8c40-413a-92a8-9dd833dbfc2c" containerName="pruner" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744667 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac60a2ae-8790-4e3d-a2c0-ba394a8140a6" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744675 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e54166e-b9e8-4136-9f06-09e8480555b7" containerName="pruner" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744682 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c3242ed-a5f8-4078-b2fc-1031ce5e92b1" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744688 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9dbdce0-d5c5-4b60-9d4b-745beca12ee8" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744695 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c9685e-2941-4778-97d2-f745955fa577" containerName="oauth-openshift" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744705 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e111a15d-a0c1-470b-9723-e4bbcc2265c1" containerName="collect-profiles" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.744715 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="400cc27a-45ec-4164-9af6-7c120fff6351" containerName="registry-server" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.745161 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.765528 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7f54ff7574-k6bww"] Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784555 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-trusted-ca-bundle\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784613 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-cliconfig\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784641 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-router-certs\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784671 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-login\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784729 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-session\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784751 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gczms\" (UniqueName: \"kubernetes.io/projected/45c9685e-2941-4778-97d2-f745955fa577-kube-api-access-gczms\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784805 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-ocp-branding-template\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784834 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-error\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784874 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/45c9685e-2941-4778-97d2-f745955fa577-audit-dir\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.784905 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-audit-policies\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.785021 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-provider-selection\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.785068 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-idp-0-file-data\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.785097 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-serving-cert\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.785128 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-service-ca\") pod \"45c9685e-2941-4778-97d2-f745955fa577\" (UID: \"45c9685e-2941-4778-97d2-f745955fa577\") " Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.787077 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-session\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.787251 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.787424 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.787704 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-template-login\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.787787 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.787879 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.787974 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.788206 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-audit-policies\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.788281 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/30b02da5-e6aa-4ea1-bc38-21f7577e316f-audit-dir\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.788362 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.788434 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.788519 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-template-error\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.788576 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsggk\" (UniqueName: \"kubernetes.io/projected/30b02da5-e6aa-4ea1-bc38-21f7577e316f-kube-api-access-gsggk\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.788615 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.789915 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.790452 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.792356 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45c9685e-2941-4778-97d2-f745955fa577-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.793915 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.795261 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.796745 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.796874 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.797300 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.797505 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.797916 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45c9685e-2941-4778-97d2-f745955fa577-kube-api-access-gczms" (OuterVolumeSpecName: "kube-api-access-gczms") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "kube-api-access-gczms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.811212 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.811864 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.812179 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.812396 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "45c9685e-2941-4778-97d2-f745955fa577" (UID: "45c9685e-2941-4778-97d2-f745955fa577"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.889965 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890038 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890065 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-audit-policies\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890157 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/30b02da5-e6aa-4ea1-bc38-21f7577e316f-audit-dir\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890203 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890228 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890271 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-template-error\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890298 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsggk\" (UniqueName: \"kubernetes.io/projected/30b02da5-e6aa-4ea1-bc38-21f7577e316f-kube-api-access-gsggk\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890322 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890344 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/30b02da5-e6aa-4ea1-bc38-21f7577e316f-audit-dir\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890386 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-session\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890561 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890633 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890782 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-template-login\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890817 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890955 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890976 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.890995 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891032 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891050 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891061 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891072 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891083 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891095 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gczms\" (UniqueName: \"kubernetes.io/projected/45c9685e-2941-4778-97d2-f745955fa577-kube-api-access-gczms\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891106 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891120 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891133 4770 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/45c9685e-2941-4778-97d2-f745955fa577-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891147 4770 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/45c9685e-2941-4778-97d2-f745955fa577-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891163 4770 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/45c9685e-2941-4778-97d2-f745955fa577-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891227 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-audit-policies\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.891576 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.892117 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.892692 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.894339 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.894498 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.895145 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-template-error\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.896461 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.896830 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-session\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.898063 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-user-template-login\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.898587 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.900128 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/30b02da5-e6aa-4ea1-bc38-21f7577e316f-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:09 crc kubenswrapper[4770]: I1004 03:07:09.908785 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsggk\" (UniqueName: \"kubernetes.io/projected/30b02da5-e6aa-4ea1-bc38-21f7577e316f-kube-api-access-gsggk\") pod \"oauth-openshift-7f54ff7574-k6bww\" (UID: \"30b02da5-e6aa-4ea1-bc38-21f7577e316f\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:10 crc kubenswrapper[4770]: I1004 03:07:10.068392 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:10 crc kubenswrapper[4770]: I1004 03:07:10.337349 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7f54ff7574-k6bww"] Oct 04 03:07:10 crc kubenswrapper[4770]: I1004 03:07:10.483599 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" event={"ID":"30b02da5-e6aa-4ea1-bc38-21f7577e316f","Type":"ContainerStarted","Data":"1acc7f831dfda4a313a7c7890c33b4d3df20bcd0017266ee3eee2a514a8ce49b"} Oct 04 03:07:10 crc kubenswrapper[4770]: I1004 03:07:10.485455 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" event={"ID":"45c9685e-2941-4778-97d2-f745955fa577","Type":"ContainerDied","Data":"15b7814bdacbb602144f70721f0934dbb8bb99dc3511897a7625906f6e00b5bf"} Oct 04 03:07:10 crc kubenswrapper[4770]: I1004 03:07:10.485522 4770 scope.go:117] "RemoveContainer" containerID="dd0d47a0ef811a3f1a95564d6f288e0a3dff34b5e663f02dadeeb9aa0a38d763" Oct 04 03:07:10 crc kubenswrapper[4770]: I1004 03:07:10.485678 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6r8rk" Oct 04 03:07:10 crc kubenswrapper[4770]: I1004 03:07:10.535103 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r8rk"] Oct 04 03:07:10 crc kubenswrapper[4770]: I1004 03:07:10.537127 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6r8rk"] Oct 04 03:07:11 crc kubenswrapper[4770]: I1004 03:07:11.498110 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" event={"ID":"30b02da5-e6aa-4ea1-bc38-21f7577e316f","Type":"ContainerStarted","Data":"23070e2c5fe0eb16ea24c00f21fa3c7e74da0c87233632cdc92ca0a9dd839aeb"} Oct 04 03:07:11 crc kubenswrapper[4770]: I1004 03:07:11.498459 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:11 crc kubenswrapper[4770]: I1004 03:07:11.508715 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" Oct 04 03:07:11 crc kubenswrapper[4770]: I1004 03:07:11.528802 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7f54ff7574-k6bww" podStartSLOduration=27.528771979 podStartE2EDuration="27.528771979s" podCreationTimestamp="2025-10-04 03:06:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:07:11.522878883 +0000 UTC m=+242.814888635" watchObservedRunningTime="2025-10-04 03:07:11.528771979 +0000 UTC m=+242.820781731" Oct 04 03:07:11 crc kubenswrapper[4770]: I1004 03:07:11.682222 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45c9685e-2941-4778-97d2-f745955fa577" path="/var/lib/kubelet/pods/45c9685e-2941-4778-97d2-f745955fa577/volumes" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.274778 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m4jht"] Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.275833 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m4jht" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerName="registry-server" containerID="cri-o://0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62" gracePeriod=30 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.294824 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dcs7t"] Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.295589 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dcs7t" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerName="registry-server" containerID="cri-o://b73c0d574135fb618dec4515ea9891db47ad4a2e00d7ccbcb15640ede0599544" gracePeriod=30 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.311063 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2fqvm"] Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.311587 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" podUID="dac89ce4-6624-4d5b-a00f-5c6ffa342535" containerName="marketplace-operator" containerID="cri-o://ef2a0ac9704bf0d07848cc17fff8d04787b9b983797b6f601eabef4c2a58d51e" gracePeriod=30 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.324478 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9zkn"] Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.324800 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n9zkn" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerName="registry-server" containerID="cri-o://741c73b49d6e423d55caca41271806e8ccd46dfb5ac23bf8431dadf17e3f6973" gracePeriod=30 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.333050 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pvjjj"] Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.333490 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pvjjj" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerName="registry-server" containerID="cri-o://cd332699ab2404ce4b67af9670fd442e8c91c75bee97f6455a41444f46ac1361" gracePeriod=30 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.352124 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-46c6n"] Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.353223 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.359713 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-46c6n"] Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.450477 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/080999d9-34f8-445c-b305-e41097107afc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-46c6n\" (UID: \"080999d9-34f8-445c-b305-e41097107afc\") " pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.450544 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqz52\" (UniqueName: \"kubernetes.io/projected/080999d9-34f8-445c-b305-e41097107afc-kube-api-access-jqz52\") pod \"marketplace-operator-79b997595-46c6n\" (UID: \"080999d9-34f8-445c-b305-e41097107afc\") " pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.450575 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/080999d9-34f8-445c-b305-e41097107afc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-46c6n\" (UID: \"080999d9-34f8-445c-b305-e41097107afc\") " pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: E1004 03:07:39.511456 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62 is running failed: container process not found" containerID="0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 03:07:39 crc kubenswrapper[4770]: E1004 03:07:39.515372 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62 is running failed: container process not found" containerID="0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 03:07:39 crc kubenswrapper[4770]: E1004 03:07:39.516285 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62 is running failed: container process not found" containerID="0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 03:07:39 crc kubenswrapper[4770]: E1004 03:07:39.516361 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-m4jht" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerName="registry-server" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.552068 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/080999d9-34f8-445c-b305-e41097107afc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-46c6n\" (UID: \"080999d9-34f8-445c-b305-e41097107afc\") " pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.552125 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqz52\" (UniqueName: \"kubernetes.io/projected/080999d9-34f8-445c-b305-e41097107afc-kube-api-access-jqz52\") pod \"marketplace-operator-79b997595-46c6n\" (UID: \"080999d9-34f8-445c-b305-e41097107afc\") " pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.552147 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/080999d9-34f8-445c-b305-e41097107afc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-46c6n\" (UID: \"080999d9-34f8-445c-b305-e41097107afc\") " pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.553378 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/080999d9-34f8-445c-b305-e41097107afc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-46c6n\" (UID: \"080999d9-34f8-445c-b305-e41097107afc\") " pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.560826 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/080999d9-34f8-445c-b305-e41097107afc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-46c6n\" (UID: \"080999d9-34f8-445c-b305-e41097107afc\") " pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.572927 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqz52\" (UniqueName: \"kubernetes.io/projected/080999d9-34f8-445c-b305-e41097107afc-kube-api-access-jqz52\") pod \"marketplace-operator-79b997595-46c6n\" (UID: \"080999d9-34f8-445c-b305-e41097107afc\") " pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.676744 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.700448 4770 generic.go:334] "Generic (PLEG): container finished" podID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerID="b73c0d574135fb618dec4515ea9891db47ad4a2e00d7ccbcb15640ede0599544" exitCode=0 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.700533 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dcs7t" event={"ID":"30e80d37-aeb7-4446-852c-3a3ac2b78bb8","Type":"ContainerDied","Data":"b73c0d574135fb618dec4515ea9891db47ad4a2e00d7ccbcb15640ede0599544"} Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.703600 4770 generic.go:334] "Generic (PLEG): container finished" podID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerID="cd332699ab2404ce4b67af9670fd442e8c91c75bee97f6455a41444f46ac1361" exitCode=0 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.703657 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjjj" event={"ID":"f8207392-9683-4278-a7d5-8e6af62fbf61","Type":"ContainerDied","Data":"cd332699ab2404ce4b67af9670fd442e8c91c75bee97f6455a41444f46ac1361"} Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.706820 4770 generic.go:334] "Generic (PLEG): container finished" podID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerID="0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62" exitCode=0 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.706884 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4jht" event={"ID":"d6d87160-c0c6-4cb0-8561-13988e9c5395","Type":"ContainerDied","Data":"0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62"} Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.706913 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4jht" event={"ID":"d6d87160-c0c6-4cb0-8561-13988e9c5395","Type":"ContainerDied","Data":"d9628b2898b3c0db9cf5615a222b4ddd63da2f3c1c360b038e95ee508489ddb6"} Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.706933 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9628b2898b3c0db9cf5615a222b4ddd63da2f3c1c360b038e95ee508489ddb6" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.709700 4770 generic.go:334] "Generic (PLEG): container finished" podID="dac89ce4-6624-4d5b-a00f-5c6ffa342535" containerID="ef2a0ac9704bf0d07848cc17fff8d04787b9b983797b6f601eabef4c2a58d51e" exitCode=0 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.709756 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" event={"ID":"dac89ce4-6624-4d5b-a00f-5c6ffa342535","Type":"ContainerDied","Data":"ef2a0ac9704bf0d07848cc17fff8d04787b9b983797b6f601eabef4c2a58d51e"} Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.711957 4770 generic.go:334] "Generic (PLEG): container finished" podID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerID="741c73b49d6e423d55caca41271806e8ccd46dfb5ac23bf8431dadf17e3f6973" exitCode=0 Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.711978 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9zkn" event={"ID":"bba2dc2c-e45e-43c2-8d02-ee01bf41143a","Type":"ContainerDied","Data":"741c73b49d6e423d55caca41271806e8ccd46dfb5ac23bf8431dadf17e3f6973"} Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.743623 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.761731 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.792234 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.851636 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.856496 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s28jc\" (UniqueName: \"kubernetes.io/projected/d6d87160-c0c6-4cb0-8561-13988e9c5395-kube-api-access-s28jc\") pod \"d6d87160-c0c6-4cb0-8561-13988e9c5395\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.856608 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-trusted-ca\") pod \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.856639 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-operator-metrics\") pod \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.856712 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-utilities\") pod \"d6d87160-c0c6-4cb0-8561-13988e9c5395\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.856749 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj2jb\" (UniqueName: \"kubernetes.io/projected/dac89ce4-6624-4d5b-a00f-5c6ffa342535-kube-api-access-wj2jb\") pod \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\" (UID: \"dac89ce4-6624-4d5b-a00f-5c6ffa342535\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.856812 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-catalog-content\") pod \"d6d87160-c0c6-4cb0-8561-13988e9c5395\" (UID: \"d6d87160-c0c6-4cb0-8561-13988e9c5395\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.857613 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "dac89ce4-6624-4d5b-a00f-5c6ffa342535" (UID: "dac89ce4-6624-4d5b-a00f-5c6ffa342535"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.858940 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-utilities" (OuterVolumeSpecName: "utilities") pod "d6d87160-c0c6-4cb0-8561-13988e9c5395" (UID: "d6d87160-c0c6-4cb0-8561-13988e9c5395"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.863072 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6d87160-c0c6-4cb0-8561-13988e9c5395-kube-api-access-s28jc" (OuterVolumeSpecName: "kube-api-access-s28jc") pod "d6d87160-c0c6-4cb0-8561-13988e9c5395" (UID: "d6d87160-c0c6-4cb0-8561-13988e9c5395"). InnerVolumeSpecName "kube-api-access-s28jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.863626 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "dac89ce4-6624-4d5b-a00f-5c6ffa342535" (UID: "dac89ce4-6624-4d5b-a00f-5c6ffa342535"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.863755 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dac89ce4-6624-4d5b-a00f-5c6ffa342535-kube-api-access-wj2jb" (OuterVolumeSpecName: "kube-api-access-wj2jb") pod "dac89ce4-6624-4d5b-a00f-5c6ffa342535" (UID: "dac89ce4-6624-4d5b-a00f-5c6ffa342535"). InnerVolumeSpecName "kube-api-access-wj2jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.866410 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.940156 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6d87160-c0c6-4cb0-8561-13988e9c5395" (UID: "d6d87160-c0c6-4cb0-8561-13988e9c5395"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.957562 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp6xb\" (UniqueName: \"kubernetes.io/projected/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-kube-api-access-hp6xb\") pod \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.957627 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4fdc\" (UniqueName: \"kubernetes.io/projected/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-kube-api-access-l4fdc\") pod \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.957654 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-utilities\") pod \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.957676 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-utilities\") pod \"f8207392-9683-4278-a7d5-8e6af62fbf61\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.957821 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-utilities\") pod \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.957848 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jzwc\" (UniqueName: \"kubernetes.io/projected/f8207392-9683-4278-a7d5-8e6af62fbf61-kube-api-access-7jzwc\") pod \"f8207392-9683-4278-a7d5-8e6af62fbf61\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.957875 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-catalog-content\") pod \"f8207392-9683-4278-a7d5-8e6af62fbf61\" (UID: \"f8207392-9683-4278-a7d5-8e6af62fbf61\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.957900 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-catalog-content\") pod \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\" (UID: \"30e80d37-aeb7-4446-852c-3a3ac2b78bb8\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.957929 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-catalog-content\") pod \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\" (UID: \"bba2dc2c-e45e-43c2-8d02-ee01bf41143a\") " Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.958920 4770 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.958944 4770 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dac89ce4-6624-4d5b-a00f-5c6ffa342535-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.958958 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.958976 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj2jb\" (UniqueName: \"kubernetes.io/projected/dac89ce4-6624-4d5b-a00f-5c6ffa342535-kube-api-access-wj2jb\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.958992 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6d87160-c0c6-4cb0-8561-13988e9c5395-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.959022 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s28jc\" (UniqueName: \"kubernetes.io/projected/d6d87160-c0c6-4cb0-8561-13988e9c5395-kube-api-access-s28jc\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.964309 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-utilities" (OuterVolumeSpecName: "utilities") pod "f8207392-9683-4278-a7d5-8e6af62fbf61" (UID: "f8207392-9683-4278-a7d5-8e6af62fbf61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.964464 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-utilities" (OuterVolumeSpecName: "utilities") pod "bba2dc2c-e45e-43c2-8d02-ee01bf41143a" (UID: "bba2dc2c-e45e-43c2-8d02-ee01bf41143a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.964497 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-kube-api-access-hp6xb" (OuterVolumeSpecName: "kube-api-access-hp6xb") pod "bba2dc2c-e45e-43c2-8d02-ee01bf41143a" (UID: "bba2dc2c-e45e-43c2-8d02-ee01bf41143a"). InnerVolumeSpecName "kube-api-access-hp6xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.964895 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-utilities" (OuterVolumeSpecName: "utilities") pod "30e80d37-aeb7-4446-852c-3a3ac2b78bb8" (UID: "30e80d37-aeb7-4446-852c-3a3ac2b78bb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.968813 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8207392-9683-4278-a7d5-8e6af62fbf61-kube-api-access-7jzwc" (OuterVolumeSpecName: "kube-api-access-7jzwc") pod "f8207392-9683-4278-a7d5-8e6af62fbf61" (UID: "f8207392-9683-4278-a7d5-8e6af62fbf61"). InnerVolumeSpecName "kube-api-access-7jzwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.972107 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-kube-api-access-l4fdc" (OuterVolumeSpecName: "kube-api-access-l4fdc") pod "30e80d37-aeb7-4446-852c-3a3ac2b78bb8" (UID: "30e80d37-aeb7-4446-852c-3a3ac2b78bb8"). InnerVolumeSpecName "kube-api-access-l4fdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:07:39 crc kubenswrapper[4770]: I1004 03:07:39.973455 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bba2dc2c-e45e-43c2-8d02-ee01bf41143a" (UID: "bba2dc2c-e45e-43c2-8d02-ee01bf41143a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.027544 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30e80d37-aeb7-4446-852c-3a3ac2b78bb8" (UID: "30e80d37-aeb7-4446-852c-3a3ac2b78bb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.049373 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8207392-9683-4278-a7d5-8e6af62fbf61" (UID: "f8207392-9683-4278-a7d5-8e6af62fbf61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.061027 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.061066 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jzwc\" (UniqueName: \"kubernetes.io/projected/f8207392-9683-4278-a7d5-8e6af62fbf61-kube-api-access-7jzwc\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.061080 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.061094 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.061145 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.061160 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp6xb\" (UniqueName: \"kubernetes.io/projected/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-kube-api-access-hp6xb\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.061174 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4fdc\" (UniqueName: \"kubernetes.io/projected/30e80d37-aeb7-4446-852c-3a3ac2b78bb8-kube-api-access-l4fdc\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.061186 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bba2dc2c-e45e-43c2-8d02-ee01bf41143a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.061197 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8207392-9683-4278-a7d5-8e6af62fbf61-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.200038 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-46c6n"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.718829 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjjj" event={"ID":"f8207392-9683-4278-a7d5-8e6af62fbf61","Type":"ContainerDied","Data":"0f3fe2e8f4f45d4fc5d1974e16ae7d11236f65cdbf5818d50d234286538a4a34"} Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.720150 4770 scope.go:117] "RemoveContainer" containerID="cd332699ab2404ce4b67af9670fd442e8c91c75bee97f6455a41444f46ac1361" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.718883 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pvjjj" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.720822 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" event={"ID":"080999d9-34f8-445c-b305-e41097107afc","Type":"ContainerStarted","Data":"0fba6d3c2d53c9ed4c6795b097c9bb05d44651206933f23737d8611fc91f80a6"} Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.720862 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" event={"ID":"080999d9-34f8-445c-b305-e41097107afc","Type":"ContainerStarted","Data":"1cc316d167d7a327e5cf9251e36e8b701d61a4738370fb8a3d67b3571be1bc5a"} Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.721037 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.722559 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" event={"ID":"dac89ce4-6624-4d5b-a00f-5c6ffa342535","Type":"ContainerDied","Data":"578f810e2d8d48995228fd057cb7d33cacf8f9e9d65c1de648a8da0d4fe0b5d9"} Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.722612 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2fqvm" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.724219 4770 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-46c6n container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.724288 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" podUID="080999d9-34f8-445c-b305-e41097107afc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.727149 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9zkn" event={"ID":"bba2dc2c-e45e-43c2-8d02-ee01bf41143a","Type":"ContainerDied","Data":"7f8808ef8da203448f17c401a8e45dd322d75438cd763b08cc00cd2bbf492a89"} Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.727232 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9zkn" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.729703 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m4jht" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.729761 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dcs7t" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.729797 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dcs7t" event={"ID":"30e80d37-aeb7-4446-852c-3a3ac2b78bb8","Type":"ContainerDied","Data":"592586b37ed5f2c9bdb81150ff5c9f5f71a844ad5075c6e3af33b0b61af59679"} Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.755845 4770 scope.go:117] "RemoveContainer" containerID="e3d5d35839a4aaad2bd9775847734e6c7ff9cb9bdc3c4e45f74603e0f99a79d4" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.773138 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" podStartSLOduration=1.7731164019999999 podStartE2EDuration="1.773116402s" podCreationTimestamp="2025-10-04 03:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:07:40.772939778 +0000 UTC m=+272.064949520" watchObservedRunningTime="2025-10-04 03:07:40.773116402 +0000 UTC m=+272.065126114" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.793878 4770 scope.go:117] "RemoveContainer" containerID="307131fe5936053720116f32bfdc419f2ffd674a57db8efb2fe235020c3f0055" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.845325 4770 scope.go:117] "RemoveContainer" containerID="ef2a0ac9704bf0d07848cc17fff8d04787b9b983797b6f601eabef4c2a58d51e" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.851420 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2fqvm"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.856111 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2fqvm"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.870102 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dcs7t"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.870643 4770 scope.go:117] "RemoveContainer" containerID="741c73b49d6e423d55caca41271806e8ccd46dfb5ac23bf8431dadf17e3f6973" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.875197 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dcs7t"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.881669 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pvjjj"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.885174 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pvjjj"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.893970 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m4jht"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.902316 4770 scope.go:117] "RemoveContainer" containerID="d4ad73927db7a50979a4e55fb419a5d125d33dfe98043726424d5fc565ff2c07" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.902877 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m4jht"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.917514 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9zkn"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.920385 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9zkn"] Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.928857 4770 scope.go:117] "RemoveContainer" containerID="5cc3549c07c2da4cb528206b1e762e2847d1987f979f0546c6c83a820b78fb1a" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.941039 4770 scope.go:117] "RemoveContainer" containerID="b73c0d574135fb618dec4515ea9891db47ad4a2e00d7ccbcb15640ede0599544" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.956945 4770 scope.go:117] "RemoveContainer" containerID="0afc29f7d9fc01fdee1497b4709f6133045f26ea902e4d063a0369c48053301a" Oct 04 03:07:40 crc kubenswrapper[4770]: I1004 03:07:40.970834 4770 scope.go:117] "RemoveContainer" containerID="fc49be386e3d7e999e25a6a07b2924343cbffcd4cc36045c228be0f0f096720d" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.507457 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5h6qk"] Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.507904 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerName="extract-content" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.507947 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerName="extract-content" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.507979 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dac89ce4-6624-4d5b-a00f-5c6ffa342535" containerName="marketplace-operator" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508001 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac89ce4-6624-4d5b-a00f-5c6ffa342535" containerName="marketplace-operator" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508056 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerName="extract-utilities" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508073 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerName="extract-utilities" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508098 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerName="extract-utilities" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508116 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerName="extract-utilities" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508133 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerName="extract-utilities" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508146 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerName="extract-utilities" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508162 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerName="extract-content" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508175 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerName="extract-content" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508189 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerName="extract-content" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508201 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerName="extract-content" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508216 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508228 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508246 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508261 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508284 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerName="extract-content" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508296 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerName="extract-content" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508313 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508341 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508355 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508367 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: E1004 03:07:41.508392 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerName="extract-utilities" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508404 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerName="extract-utilities" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508559 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508578 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="dac89ce4-6624-4d5b-a00f-5c6ffa342535" containerName="marketplace-operator" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508601 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508618 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.508641 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" containerName="registry-server" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.510053 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.513040 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.517792 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5h6qk"] Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.581234 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njzf2\" (UniqueName: \"kubernetes.io/projected/10bb88a1-02bc-4344-9f59-1e4794a719a0-kube-api-access-njzf2\") pod \"certified-operators-5h6qk\" (UID: \"10bb88a1-02bc-4344-9f59-1e4794a719a0\") " pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.581289 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10bb88a1-02bc-4344-9f59-1e4794a719a0-catalog-content\") pod \"certified-operators-5h6qk\" (UID: \"10bb88a1-02bc-4344-9f59-1e4794a719a0\") " pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.581334 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10bb88a1-02bc-4344-9f59-1e4794a719a0-utilities\") pod \"certified-operators-5h6qk\" (UID: \"10bb88a1-02bc-4344-9f59-1e4794a719a0\") " pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.682794 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njzf2\" (UniqueName: \"kubernetes.io/projected/10bb88a1-02bc-4344-9f59-1e4794a719a0-kube-api-access-njzf2\") pod \"certified-operators-5h6qk\" (UID: \"10bb88a1-02bc-4344-9f59-1e4794a719a0\") " pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.683704 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10bb88a1-02bc-4344-9f59-1e4794a719a0-catalog-content\") pod \"certified-operators-5h6qk\" (UID: \"10bb88a1-02bc-4344-9f59-1e4794a719a0\") " pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.684052 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10bb88a1-02bc-4344-9f59-1e4794a719a0-utilities\") pod \"certified-operators-5h6qk\" (UID: \"10bb88a1-02bc-4344-9f59-1e4794a719a0\") " pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.684213 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10bb88a1-02bc-4344-9f59-1e4794a719a0-catalog-content\") pod \"certified-operators-5h6qk\" (UID: \"10bb88a1-02bc-4344-9f59-1e4794a719a0\") " pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.684845 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10bb88a1-02bc-4344-9f59-1e4794a719a0-utilities\") pod \"certified-operators-5h6qk\" (UID: \"10bb88a1-02bc-4344-9f59-1e4794a719a0\") " pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.699676 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30e80d37-aeb7-4446-852c-3a3ac2b78bb8" path="/var/lib/kubelet/pods/30e80d37-aeb7-4446-852c-3a3ac2b78bb8/volumes" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.700466 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bba2dc2c-e45e-43c2-8d02-ee01bf41143a" path="/var/lib/kubelet/pods/bba2dc2c-e45e-43c2-8d02-ee01bf41143a/volumes" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.701115 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6d87160-c0c6-4cb0-8561-13988e9c5395" path="/var/lib/kubelet/pods/d6d87160-c0c6-4cb0-8561-13988e9c5395/volumes" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.704634 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dac89ce4-6624-4d5b-a00f-5c6ffa342535" path="/var/lib/kubelet/pods/dac89ce4-6624-4d5b-a00f-5c6ffa342535/volumes" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.705488 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8207392-9683-4278-a7d5-8e6af62fbf61" path="/var/lib/kubelet/pods/f8207392-9683-4278-a7d5-8e6af62fbf61/volumes" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.706920 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njzf2\" (UniqueName: \"kubernetes.io/projected/10bb88a1-02bc-4344-9f59-1e4794a719a0-kube-api-access-njzf2\") pod \"certified-operators-5h6qk\" (UID: \"10bb88a1-02bc-4344-9f59-1e4794a719a0\") " pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.707077 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-slzl4"] Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.708657 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-slzl4"] Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.708768 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.710845 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.750274 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-46c6n" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.785922 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-utilities\") pod \"community-operators-slzl4\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.785998 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-catalog-content\") pod \"community-operators-slzl4\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.786315 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvkrq\" (UniqueName: \"kubernetes.io/projected/8322250a-b542-48c5-a8df-d282483515bb-kube-api-access-nvkrq\") pod \"community-operators-slzl4\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.840777 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.888087 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-utilities\") pod \"community-operators-slzl4\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.888183 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-catalog-content\") pod \"community-operators-slzl4\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.888346 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvkrq\" (UniqueName: \"kubernetes.io/projected/8322250a-b542-48c5-a8df-d282483515bb-kube-api-access-nvkrq\") pod \"community-operators-slzl4\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.888648 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-utilities\") pod \"community-operators-slzl4\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.888912 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-catalog-content\") pod \"community-operators-slzl4\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:41 crc kubenswrapper[4770]: I1004 03:07:41.910435 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvkrq\" (UniqueName: \"kubernetes.io/projected/8322250a-b542-48c5-a8df-d282483515bb-kube-api-access-nvkrq\") pod \"community-operators-slzl4\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:42 crc kubenswrapper[4770]: I1004 03:07:42.036569 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:42 crc kubenswrapper[4770]: I1004 03:07:42.056723 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5h6qk"] Oct 04 03:07:42 crc kubenswrapper[4770]: W1004 03:07:42.069758 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10bb88a1_02bc_4344_9f59_1e4794a719a0.slice/crio-38d1e59be4c3e5bad33dc820602d91a8baecb2ea5b87254890f9a20866aa42e5 WatchSource:0}: Error finding container 38d1e59be4c3e5bad33dc820602d91a8baecb2ea5b87254890f9a20866aa42e5: Status 404 returned error can't find the container with id 38d1e59be4c3e5bad33dc820602d91a8baecb2ea5b87254890f9a20866aa42e5 Oct 04 03:07:42 crc kubenswrapper[4770]: I1004 03:07:42.224489 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-slzl4"] Oct 04 03:07:42 crc kubenswrapper[4770]: W1004 03:07:42.235506 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8322250a_b542_48c5_a8df_d282483515bb.slice/crio-783fc8ca1452735636382ab4d041a6990bc31fd7dee15efc60a56b5b18ea87b1 WatchSource:0}: Error finding container 783fc8ca1452735636382ab4d041a6990bc31fd7dee15efc60a56b5b18ea87b1: Status 404 returned error can't find the container with id 783fc8ca1452735636382ab4d041a6990bc31fd7dee15efc60a56b5b18ea87b1 Oct 04 03:07:42 crc kubenswrapper[4770]: I1004 03:07:42.748849 4770 generic.go:334] "Generic (PLEG): container finished" podID="8322250a-b542-48c5-a8df-d282483515bb" containerID="f29039087d4fd9e9c9f9020bb9dcaf739ab1306a9b6fe949f2761144408bdbd1" exitCode=0 Oct 04 03:07:42 crc kubenswrapper[4770]: I1004 03:07:42.748927 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slzl4" event={"ID":"8322250a-b542-48c5-a8df-d282483515bb","Type":"ContainerDied","Data":"f29039087d4fd9e9c9f9020bb9dcaf739ab1306a9b6fe949f2761144408bdbd1"} Oct 04 03:07:42 crc kubenswrapper[4770]: I1004 03:07:42.749529 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slzl4" event={"ID":"8322250a-b542-48c5-a8df-d282483515bb","Type":"ContainerStarted","Data":"783fc8ca1452735636382ab4d041a6990bc31fd7dee15efc60a56b5b18ea87b1"} Oct 04 03:07:42 crc kubenswrapper[4770]: I1004 03:07:42.768517 4770 generic.go:334] "Generic (PLEG): container finished" podID="10bb88a1-02bc-4344-9f59-1e4794a719a0" containerID="2ec93ce0f448461f9e153dcfbbd380012bda09f1bde0cab0960d28794da665e5" exitCode=0 Oct 04 03:07:42 crc kubenswrapper[4770]: I1004 03:07:42.768636 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5h6qk" event={"ID":"10bb88a1-02bc-4344-9f59-1e4794a719a0","Type":"ContainerDied","Data":"2ec93ce0f448461f9e153dcfbbd380012bda09f1bde0cab0960d28794da665e5"} Oct 04 03:07:42 crc kubenswrapper[4770]: I1004 03:07:42.768729 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5h6qk" event={"ID":"10bb88a1-02bc-4344-9f59-1e4794a719a0","Type":"ContainerStarted","Data":"38d1e59be4c3e5bad33dc820602d91a8baecb2ea5b87254890f9a20866aa42e5"} Oct 04 03:07:43 crc kubenswrapper[4770]: I1004 03:07:43.897325 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ppnbx"] Oct 04 03:07:43 crc kubenswrapper[4770]: I1004 03:07:43.899088 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:43 crc kubenswrapper[4770]: I1004 03:07:43.902381 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 03:07:43 crc kubenswrapper[4770]: I1004 03:07:43.908238 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ppnbx"] Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.019114 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14b26c94-8153-4918-afe0-098a8e16a725-catalog-content\") pod \"redhat-marketplace-ppnbx\" (UID: \"14b26c94-8153-4918-afe0-098a8e16a725\") " pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.019227 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14b26c94-8153-4918-afe0-098a8e16a725-utilities\") pod \"redhat-marketplace-ppnbx\" (UID: \"14b26c94-8153-4918-afe0-098a8e16a725\") " pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.019279 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx7kp\" (UniqueName: \"kubernetes.io/projected/14b26c94-8153-4918-afe0-098a8e16a725-kube-api-access-lx7kp\") pod \"redhat-marketplace-ppnbx\" (UID: \"14b26c94-8153-4918-afe0-098a8e16a725\") " pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.089206 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mvtk8"] Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.090291 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.093512 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.106928 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mvtk8"] Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.120452 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14b26c94-8153-4918-afe0-098a8e16a725-catalog-content\") pod \"redhat-marketplace-ppnbx\" (UID: \"14b26c94-8153-4918-afe0-098a8e16a725\") " pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.120498 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/071fc5ae-cc19-4618-aa7f-53518791bb22-utilities\") pod \"redhat-operators-mvtk8\" (UID: \"071fc5ae-cc19-4618-aa7f-53518791bb22\") " pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.120529 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14b26c94-8153-4918-afe0-098a8e16a725-utilities\") pod \"redhat-marketplace-ppnbx\" (UID: \"14b26c94-8153-4918-afe0-098a8e16a725\") " pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.120552 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx7kp\" (UniqueName: \"kubernetes.io/projected/14b26c94-8153-4918-afe0-098a8e16a725-kube-api-access-lx7kp\") pod \"redhat-marketplace-ppnbx\" (UID: \"14b26c94-8153-4918-afe0-098a8e16a725\") " pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.120613 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/071fc5ae-cc19-4618-aa7f-53518791bb22-catalog-content\") pod \"redhat-operators-mvtk8\" (UID: \"071fc5ae-cc19-4618-aa7f-53518791bb22\") " pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.120629 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2dfq\" (UniqueName: \"kubernetes.io/projected/071fc5ae-cc19-4618-aa7f-53518791bb22-kube-api-access-r2dfq\") pod \"redhat-operators-mvtk8\" (UID: \"071fc5ae-cc19-4618-aa7f-53518791bb22\") " pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.120972 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14b26c94-8153-4918-afe0-098a8e16a725-catalog-content\") pod \"redhat-marketplace-ppnbx\" (UID: \"14b26c94-8153-4918-afe0-098a8e16a725\") " pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.121168 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14b26c94-8153-4918-afe0-098a8e16a725-utilities\") pod \"redhat-marketplace-ppnbx\" (UID: \"14b26c94-8153-4918-afe0-098a8e16a725\") " pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.142968 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx7kp\" (UniqueName: \"kubernetes.io/projected/14b26c94-8153-4918-afe0-098a8e16a725-kube-api-access-lx7kp\") pod \"redhat-marketplace-ppnbx\" (UID: \"14b26c94-8153-4918-afe0-098a8e16a725\") " pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.222300 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/071fc5ae-cc19-4618-aa7f-53518791bb22-catalog-content\") pod \"redhat-operators-mvtk8\" (UID: \"071fc5ae-cc19-4618-aa7f-53518791bb22\") " pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.222388 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2dfq\" (UniqueName: \"kubernetes.io/projected/071fc5ae-cc19-4618-aa7f-53518791bb22-kube-api-access-r2dfq\") pod \"redhat-operators-mvtk8\" (UID: \"071fc5ae-cc19-4618-aa7f-53518791bb22\") " pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.222440 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/071fc5ae-cc19-4618-aa7f-53518791bb22-utilities\") pod \"redhat-operators-mvtk8\" (UID: \"071fc5ae-cc19-4618-aa7f-53518791bb22\") " pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.223184 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/071fc5ae-cc19-4618-aa7f-53518791bb22-catalog-content\") pod \"redhat-operators-mvtk8\" (UID: \"071fc5ae-cc19-4618-aa7f-53518791bb22\") " pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.223692 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/071fc5ae-cc19-4618-aa7f-53518791bb22-utilities\") pod \"redhat-operators-mvtk8\" (UID: \"071fc5ae-cc19-4618-aa7f-53518791bb22\") " pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.246984 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2dfq\" (UniqueName: \"kubernetes.io/projected/071fc5ae-cc19-4618-aa7f-53518791bb22-kube-api-access-r2dfq\") pod \"redhat-operators-mvtk8\" (UID: \"071fc5ae-cc19-4618-aa7f-53518791bb22\") " pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.269063 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.409378 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.455521 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ppnbx"] Oct 04 03:07:44 crc kubenswrapper[4770]: W1004 03:07:44.462498 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14b26c94_8153_4918_afe0_098a8e16a725.slice/crio-6e2b87433acd08de36e83489325a5944a04f7447d2a9cb4c96ca65a73d48f7f0 WatchSource:0}: Error finding container 6e2b87433acd08de36e83489325a5944a04f7447d2a9cb4c96ca65a73d48f7f0: Status 404 returned error can't find the container with id 6e2b87433acd08de36e83489325a5944a04f7447d2a9cb4c96ca65a73d48f7f0 Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.624158 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mvtk8"] Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.784648 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvtk8" event={"ID":"071fc5ae-cc19-4618-aa7f-53518791bb22","Type":"ContainerStarted","Data":"b30603da3181750c06448f052636defd6e53597021bdd33562d0b335261d9fd4"} Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.786501 4770 generic.go:334] "Generic (PLEG): container finished" podID="14b26c94-8153-4918-afe0-098a8e16a725" containerID="111ffc42769447bc8fd3e0b08513a8bd56bfd439fbb1a31c5fcedf15653542d0" exitCode=0 Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.786549 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ppnbx" event={"ID":"14b26c94-8153-4918-afe0-098a8e16a725","Type":"ContainerDied","Data":"111ffc42769447bc8fd3e0b08513a8bd56bfd439fbb1a31c5fcedf15653542d0"} Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.786569 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ppnbx" event={"ID":"14b26c94-8153-4918-afe0-098a8e16a725","Type":"ContainerStarted","Data":"6e2b87433acd08de36e83489325a5944a04f7447d2a9cb4c96ca65a73d48f7f0"} Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.791870 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slzl4" event={"ID":"8322250a-b542-48c5-a8df-d282483515bb","Type":"ContainerDied","Data":"f741c5e1637129a3f7bf342e9bb3f494d84f14838e4da28cddef19b8c6fe1c1e"} Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.791563 4770 generic.go:334] "Generic (PLEG): container finished" podID="8322250a-b542-48c5-a8df-d282483515bb" containerID="f741c5e1637129a3f7bf342e9bb3f494d84f14838e4da28cddef19b8c6fe1c1e" exitCode=0 Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.802197 4770 generic.go:334] "Generic (PLEG): container finished" podID="10bb88a1-02bc-4344-9f59-1e4794a719a0" containerID="a0c421ddafcdd52c53ed1d0f9b77dfaefa7b789441580b62f2d77ba2624203ca" exitCode=0 Oct 04 03:07:44 crc kubenswrapper[4770]: I1004 03:07:44.802242 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5h6qk" event={"ID":"10bb88a1-02bc-4344-9f59-1e4794a719a0","Type":"ContainerDied","Data":"a0c421ddafcdd52c53ed1d0f9b77dfaefa7b789441580b62f2d77ba2624203ca"} Oct 04 03:07:45 crc kubenswrapper[4770]: I1004 03:07:45.810913 4770 generic.go:334] "Generic (PLEG): container finished" podID="071fc5ae-cc19-4618-aa7f-53518791bb22" containerID="a38101a87c54462476650c55d2418939d8400b6e267a6980b92556cf51fa296f" exitCode=0 Oct 04 03:07:45 crc kubenswrapper[4770]: I1004 03:07:45.811166 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvtk8" event={"ID":"071fc5ae-cc19-4618-aa7f-53518791bb22","Type":"ContainerDied","Data":"a38101a87c54462476650c55d2418939d8400b6e267a6980b92556cf51fa296f"} Oct 04 03:07:45 crc kubenswrapper[4770]: I1004 03:07:45.816903 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slzl4" event={"ID":"8322250a-b542-48c5-a8df-d282483515bb","Type":"ContainerStarted","Data":"7d9e936ef71a3e3df3718432cdb01f0d59c56448206548422a1758fde37cf89d"} Oct 04 03:07:45 crc kubenswrapper[4770]: I1004 03:07:45.820259 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5h6qk" event={"ID":"10bb88a1-02bc-4344-9f59-1e4794a719a0","Type":"ContainerStarted","Data":"e413afd25bad131520008972ef6709746cfde8e0b32f905f01600b629aa00474"} Oct 04 03:07:45 crc kubenswrapper[4770]: I1004 03:07:45.855764 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5h6qk" podStartSLOduration=2.194384528 podStartE2EDuration="4.855743284s" podCreationTimestamp="2025-10-04 03:07:41 +0000 UTC" firstStartedPulling="2025-10-04 03:07:42.770954934 +0000 UTC m=+274.062964646" lastFinishedPulling="2025-10-04 03:07:45.43231369 +0000 UTC m=+276.724323402" observedRunningTime="2025-10-04 03:07:45.853040985 +0000 UTC m=+277.145050707" watchObservedRunningTime="2025-10-04 03:07:45.855743284 +0000 UTC m=+277.147752996" Oct 04 03:07:45 crc kubenswrapper[4770]: I1004 03:07:45.881857 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-slzl4" podStartSLOduration=2.455200517 podStartE2EDuration="4.881839189s" podCreationTimestamp="2025-10-04 03:07:41 +0000 UTC" firstStartedPulling="2025-10-04 03:07:42.758635109 +0000 UTC m=+274.050644821" lastFinishedPulling="2025-10-04 03:07:45.185273781 +0000 UTC m=+276.477283493" observedRunningTime="2025-10-04 03:07:45.878774331 +0000 UTC m=+277.170784043" watchObservedRunningTime="2025-10-04 03:07:45.881839189 +0000 UTC m=+277.173848901" Oct 04 03:07:46 crc kubenswrapper[4770]: I1004 03:07:46.828987 4770 generic.go:334] "Generic (PLEG): container finished" podID="14b26c94-8153-4918-afe0-098a8e16a725" containerID="ca41554f2fe911fc9a60b4c7f43a796f1d830b1af7fb419602924890b7372d46" exitCode=0 Oct 04 03:07:46 crc kubenswrapper[4770]: I1004 03:07:46.829054 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ppnbx" event={"ID":"14b26c94-8153-4918-afe0-098a8e16a725","Type":"ContainerDied","Data":"ca41554f2fe911fc9a60b4c7f43a796f1d830b1af7fb419602924890b7372d46"} Oct 04 03:07:48 crc kubenswrapper[4770]: I1004 03:07:48.842342 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvtk8" event={"ID":"071fc5ae-cc19-4618-aa7f-53518791bb22","Type":"ContainerStarted","Data":"fadaa9b12ae00d8285f86ffd8d4177ded23e7edbf99c1a9c8747fc52541bfd84"} Oct 04 03:07:48 crc kubenswrapper[4770]: I1004 03:07:48.845286 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ppnbx" event={"ID":"14b26c94-8153-4918-afe0-098a8e16a725","Type":"ContainerStarted","Data":"214f219ba926dcf13562e0cdbed2537a8a9db1c5ddd93682433e19483eef5a75"} Oct 04 03:07:48 crc kubenswrapper[4770]: I1004 03:07:48.880222 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ppnbx" podStartSLOduration=3.421765534 podStartE2EDuration="5.880195306s" podCreationTimestamp="2025-10-04 03:07:43 +0000 UTC" firstStartedPulling="2025-10-04 03:07:44.788131548 +0000 UTC m=+276.080141260" lastFinishedPulling="2025-10-04 03:07:47.24656131 +0000 UTC m=+278.538571032" observedRunningTime="2025-10-04 03:07:48.875180929 +0000 UTC m=+280.167190641" watchObservedRunningTime="2025-10-04 03:07:48.880195306 +0000 UTC m=+280.172205038" Oct 04 03:07:49 crc kubenswrapper[4770]: I1004 03:07:49.853751 4770 generic.go:334] "Generic (PLEG): container finished" podID="071fc5ae-cc19-4618-aa7f-53518791bb22" containerID="fadaa9b12ae00d8285f86ffd8d4177ded23e7edbf99c1a9c8747fc52541bfd84" exitCode=0 Oct 04 03:07:49 crc kubenswrapper[4770]: I1004 03:07:49.853876 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvtk8" event={"ID":"071fc5ae-cc19-4618-aa7f-53518791bb22","Type":"ContainerDied","Data":"fadaa9b12ae00d8285f86ffd8d4177ded23e7edbf99c1a9c8747fc52541bfd84"} Oct 04 03:07:51 crc kubenswrapper[4770]: I1004 03:07:51.841310 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:51 crc kubenswrapper[4770]: I1004 03:07:51.841525 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:51 crc kubenswrapper[4770]: I1004 03:07:51.870322 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvtk8" event={"ID":"071fc5ae-cc19-4618-aa7f-53518791bb22","Type":"ContainerStarted","Data":"e937fbfb86a282fc2c0ba5be4032af99dd0b50dcbeb445cfd7f4077661d8c836"} Oct 04 03:07:51 crc kubenswrapper[4770]: I1004 03:07:51.897705 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mvtk8" podStartSLOduration=2.9712411210000003 podStartE2EDuration="7.89768555s" podCreationTimestamp="2025-10-04 03:07:44 +0000 UTC" firstStartedPulling="2025-10-04 03:07:45.81321792 +0000 UTC m=+277.105227632" lastFinishedPulling="2025-10-04 03:07:50.739662319 +0000 UTC m=+282.031672061" observedRunningTime="2025-10-04 03:07:51.896022358 +0000 UTC m=+283.188032070" watchObservedRunningTime="2025-10-04 03:07:51.89768555 +0000 UTC m=+283.189695262" Oct 04 03:07:51 crc kubenswrapper[4770]: I1004 03:07:51.903335 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:51 crc kubenswrapper[4770]: I1004 03:07:51.950279 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5h6qk" Oct 04 03:07:52 crc kubenswrapper[4770]: I1004 03:07:52.037391 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:52 crc kubenswrapper[4770]: I1004 03:07:52.037776 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:52 crc kubenswrapper[4770]: I1004 03:07:52.076772 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:52 crc kubenswrapper[4770]: I1004 03:07:52.920361 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-slzl4" Oct 04 03:07:54 crc kubenswrapper[4770]: I1004 03:07:54.269197 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:54 crc kubenswrapper[4770]: I1004 03:07:54.269573 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:54 crc kubenswrapper[4770]: I1004 03:07:54.328545 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:54 crc kubenswrapper[4770]: I1004 03:07:54.410288 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:54 crc kubenswrapper[4770]: I1004 03:07:54.410371 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:07:54 crc kubenswrapper[4770]: I1004 03:07:54.923632 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ppnbx" Oct 04 03:07:55 crc kubenswrapper[4770]: I1004 03:07:55.457981 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mvtk8" podUID="071fc5ae-cc19-4618-aa7f-53518791bb22" containerName="registry-server" probeResult="failure" output=< Oct 04 03:07:55 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 03:07:55 crc kubenswrapper[4770]: > Oct 04 03:08:04 crc kubenswrapper[4770]: I1004 03:08:04.487632 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:08:04 crc kubenswrapper[4770]: I1004 03:08:04.556572 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mvtk8" Oct 04 03:09:31 crc kubenswrapper[4770]: I1004 03:09:31.796432 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:09:31 crc kubenswrapper[4770]: I1004 03:09:31.797245 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.295478 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nq6v7"] Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.297792 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.324161 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nq6v7"] Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.487973 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1cefe9c4-657b-442c-b2b6-4e643d873fdc-registry-certificates\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.488050 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1cefe9c4-657b-442c-b2b6-4e643d873fdc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.488075 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1cefe9c4-657b-442c-b2b6-4e643d873fdc-registry-tls\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.488099 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1cefe9c4-657b-442c-b2b6-4e643d873fdc-bound-sa-token\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.488143 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.488300 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1cefe9c4-657b-442c-b2b6-4e643d873fdc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.488412 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1cefe9c4-657b-442c-b2b6-4e643d873fdc-trusted-ca\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.488453 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd48j\" (UniqueName: \"kubernetes.io/projected/1cefe9c4-657b-442c-b2b6-4e643d873fdc-kube-api-access-hd48j\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.532321 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.592662 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1cefe9c4-657b-442c-b2b6-4e643d873fdc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.592776 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1cefe9c4-657b-442c-b2b6-4e643d873fdc-registry-tls\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.592883 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1cefe9c4-657b-442c-b2b6-4e643d873fdc-bound-sa-token\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.592972 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1cefe9c4-657b-442c-b2b6-4e643d873fdc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.593042 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1cefe9c4-657b-442c-b2b6-4e643d873fdc-trusted-ca\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.593089 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd48j\" (UniqueName: \"kubernetes.io/projected/1cefe9c4-657b-442c-b2b6-4e643d873fdc-kube-api-access-hd48j\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.593174 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1cefe9c4-657b-442c-b2b6-4e643d873fdc-registry-certificates\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.595390 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1cefe9c4-657b-442c-b2b6-4e643d873fdc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.595767 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1cefe9c4-657b-442c-b2b6-4e643d873fdc-trusted-ca\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.603347 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1cefe9c4-657b-442c-b2b6-4e643d873fdc-registry-tls\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.603784 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1cefe9c4-657b-442c-b2b6-4e643d873fdc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.604524 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1cefe9c4-657b-442c-b2b6-4e643d873fdc-registry-certificates\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.619326 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1cefe9c4-657b-442c-b2b6-4e643d873fdc-bound-sa-token\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.630529 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd48j\" (UniqueName: \"kubernetes.io/projected/1cefe9c4-657b-442c-b2b6-4e643d873fdc-kube-api-access-hd48j\") pod \"image-registry-66df7c8f76-nq6v7\" (UID: \"1cefe9c4-657b-442c-b2b6-4e643d873fdc\") " pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:41 crc kubenswrapper[4770]: I1004 03:09:41.921752 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:42 crc kubenswrapper[4770]: I1004 03:09:42.146642 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nq6v7"] Oct 04 03:09:42 crc kubenswrapper[4770]: I1004 03:09:42.697212 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" event={"ID":"1cefe9c4-657b-442c-b2b6-4e643d873fdc","Type":"ContainerStarted","Data":"42bdaaa00d64a6c23cb25060f0f477c5c43a75521aef359bef1d84bf1b748b21"} Oct 04 03:09:42 crc kubenswrapper[4770]: I1004 03:09:42.697669 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:09:42 crc kubenswrapper[4770]: I1004 03:09:42.697685 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" event={"ID":"1cefe9c4-657b-442c-b2b6-4e643d873fdc","Type":"ContainerStarted","Data":"22c51aa320f104794ca781d9af50d3809fc31fb2623b5394d00ba2a72161f866"} Oct 04 03:09:42 crc kubenswrapper[4770]: I1004 03:09:42.720286 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" podStartSLOduration=1.720267534 podStartE2EDuration="1.720267534s" podCreationTimestamp="2025-10-04 03:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:09:42.719430353 +0000 UTC m=+394.011440075" watchObservedRunningTime="2025-10-04 03:09:42.720267534 +0000 UTC m=+394.012277256" Oct 04 03:10:01 crc kubenswrapper[4770]: I1004 03:10:01.796134 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:10:01 crc kubenswrapper[4770]: I1004 03:10:01.797626 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:10:01 crc kubenswrapper[4770]: I1004 03:10:01.934267 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-nq6v7" Oct 04 03:10:02 crc kubenswrapper[4770]: I1004 03:10:02.003456 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lnhsd"] Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.069585 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" podUID="108c220a-bc76-4e17-884d-80f2a5b63357" containerName="registry" containerID="cri-o://72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81" gracePeriod=30 Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.441146 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.545768 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-registry-tls\") pod \"108c220a-bc76-4e17-884d-80f2a5b63357\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.545879 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9dmn\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-kube-api-access-j9dmn\") pod \"108c220a-bc76-4e17-884d-80f2a5b63357\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.546274 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"108c220a-bc76-4e17-884d-80f2a5b63357\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.546334 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/108c220a-bc76-4e17-884d-80f2a5b63357-ca-trust-extracted\") pod \"108c220a-bc76-4e17-884d-80f2a5b63357\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.546363 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-trusted-ca\") pod \"108c220a-bc76-4e17-884d-80f2a5b63357\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.546505 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/108c220a-bc76-4e17-884d-80f2a5b63357-installation-pull-secrets\") pod \"108c220a-bc76-4e17-884d-80f2a5b63357\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.546580 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-registry-certificates\") pod \"108c220a-bc76-4e17-884d-80f2a5b63357\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.546602 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-bound-sa-token\") pod \"108c220a-bc76-4e17-884d-80f2a5b63357\" (UID: \"108c220a-bc76-4e17-884d-80f2a5b63357\") " Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.547642 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "108c220a-bc76-4e17-884d-80f2a5b63357" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.547750 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "108c220a-bc76-4e17-884d-80f2a5b63357" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.553896 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "108c220a-bc76-4e17-884d-80f2a5b63357" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.554643 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "108c220a-bc76-4e17-884d-80f2a5b63357" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.555103 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-kube-api-access-j9dmn" (OuterVolumeSpecName: "kube-api-access-j9dmn") pod "108c220a-bc76-4e17-884d-80f2a5b63357" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357"). InnerVolumeSpecName "kube-api-access-j9dmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.565221 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/108c220a-bc76-4e17-884d-80f2a5b63357-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "108c220a-bc76-4e17-884d-80f2a5b63357" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.570052 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/108c220a-bc76-4e17-884d-80f2a5b63357-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "108c220a-bc76-4e17-884d-80f2a5b63357" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.575690 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "108c220a-bc76-4e17-884d-80f2a5b63357" (UID: "108c220a-bc76-4e17-884d-80f2a5b63357"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.647847 4770 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/108c220a-bc76-4e17-884d-80f2a5b63357-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.647900 4770 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.647917 4770 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.647928 4770 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.647943 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9dmn\" (UniqueName: \"kubernetes.io/projected/108c220a-bc76-4e17-884d-80f2a5b63357-kube-api-access-j9dmn\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.647954 4770 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/108c220a-bc76-4e17-884d-80f2a5b63357-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:27 crc kubenswrapper[4770]: I1004 03:10:27.647967 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/108c220a-bc76-4e17-884d-80f2a5b63357-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:28 crc kubenswrapper[4770]: I1004 03:10:28.034786 4770 generic.go:334] "Generic (PLEG): container finished" podID="108c220a-bc76-4e17-884d-80f2a5b63357" containerID="72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81" exitCode=0 Oct 04 03:10:28 crc kubenswrapper[4770]: I1004 03:10:28.034863 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" event={"ID":"108c220a-bc76-4e17-884d-80f2a5b63357","Type":"ContainerDied","Data":"72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81"} Oct 04 03:10:28 crc kubenswrapper[4770]: I1004 03:10:28.034942 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" event={"ID":"108c220a-bc76-4e17-884d-80f2a5b63357","Type":"ContainerDied","Data":"ac09181def512480c2128ccbeea7dd6d5bebc64b6e45f6196bb5ccb1b44efadc"} Oct 04 03:10:28 crc kubenswrapper[4770]: I1004 03:10:28.034973 4770 scope.go:117] "RemoveContainer" containerID="72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81" Oct 04 03:10:28 crc kubenswrapper[4770]: I1004 03:10:28.034880 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-lnhsd" Oct 04 03:10:28 crc kubenswrapper[4770]: I1004 03:10:28.068083 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lnhsd"] Oct 04 03:10:28 crc kubenswrapper[4770]: I1004 03:10:28.069695 4770 scope.go:117] "RemoveContainer" containerID="72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81" Oct 04 03:10:28 crc kubenswrapper[4770]: E1004 03:10:28.070497 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81\": container with ID starting with 72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81 not found: ID does not exist" containerID="72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81" Oct 04 03:10:28 crc kubenswrapper[4770]: I1004 03:10:28.070553 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81"} err="failed to get container status \"72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81\": rpc error: code = NotFound desc = could not find container \"72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81\": container with ID starting with 72612e94608171764c87c273fefc8b1aaf226daeb32f32b72e6d7f5ca1b62d81 not found: ID does not exist" Oct 04 03:10:28 crc kubenswrapper[4770]: I1004 03:10:28.074244 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-lnhsd"] Oct 04 03:10:29 crc kubenswrapper[4770]: I1004 03:10:29.685308 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="108c220a-bc76-4e17-884d-80f2a5b63357" path="/var/lib/kubelet/pods/108c220a-bc76-4e17-884d-80f2a5b63357/volumes" Oct 04 03:10:31 crc kubenswrapper[4770]: I1004 03:10:31.795873 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:10:31 crc kubenswrapper[4770]: I1004 03:10:31.795993 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:10:31 crc kubenswrapper[4770]: I1004 03:10:31.796103 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:10:31 crc kubenswrapper[4770]: I1004 03:10:31.797030 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"52bc3793e50aecbadbef1d828c974de853af6a75a471a3fa3860ec13be9d7c91"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:10:31 crc kubenswrapper[4770]: I1004 03:10:31.797110 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://52bc3793e50aecbadbef1d828c974de853af6a75a471a3fa3860ec13be9d7c91" gracePeriod=600 Oct 04 03:10:32 crc kubenswrapper[4770]: I1004 03:10:32.075901 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="52bc3793e50aecbadbef1d828c974de853af6a75a471a3fa3860ec13be9d7c91" exitCode=0 Oct 04 03:10:32 crc kubenswrapper[4770]: I1004 03:10:32.075961 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"52bc3793e50aecbadbef1d828c974de853af6a75a471a3fa3860ec13be9d7c91"} Oct 04 03:10:32 crc kubenswrapper[4770]: I1004 03:10:32.076036 4770 scope.go:117] "RemoveContainer" containerID="34d239b1de1b9a56bf65c8bd87b721a31e4fb97bc2d8f2a7f80c056716a830d6" Oct 04 03:10:33 crc kubenswrapper[4770]: I1004 03:10:33.085767 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"d700d3adbc22ca8f48ded121cdd0b6720fb293c613b3f9c60d656bef6938d1c8"} Oct 04 03:12:09 crc kubenswrapper[4770]: I1004 03:12:09.870681 4770 scope.go:117] "RemoveContainer" containerID="1d6ed5cbe64527c2ef27923e626e083919ee17042ce5caa5a31dbfc1a71b3d60" Oct 04 03:13:01 crc kubenswrapper[4770]: I1004 03:13:01.796272 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:13:01 crc kubenswrapper[4770]: I1004 03:13:01.797199 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:13:09 crc kubenswrapper[4770]: I1004 03:13:09.913290 4770 scope.go:117] "RemoveContainer" containerID="49a6c4b91ce5a1f4247a4216fe641c4f1aef8a4eb6025ddd89e98086b0c6446f" Oct 04 03:13:09 crc kubenswrapper[4770]: I1004 03:13:09.953451 4770 scope.go:117] "RemoveContainer" containerID="0529080b7fd73b449ca28d3af17d615d9ce2f30a3c9be9362eee05f60f33ad62" Oct 04 03:13:31 crc kubenswrapper[4770]: I1004 03:13:31.795989 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:13:31 crc kubenswrapper[4770]: I1004 03:13:31.796763 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:14:01 crc kubenswrapper[4770]: I1004 03:14:01.796264 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:14:01 crc kubenswrapper[4770]: I1004 03:14:01.798137 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:14:01 crc kubenswrapper[4770]: I1004 03:14:01.798465 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:14:01 crc kubenswrapper[4770]: I1004 03:14:01.799366 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d700d3adbc22ca8f48ded121cdd0b6720fb293c613b3f9c60d656bef6938d1c8"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:14:01 crc kubenswrapper[4770]: I1004 03:14:01.799638 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://d700d3adbc22ca8f48ded121cdd0b6720fb293c613b3f9c60d656bef6938d1c8" gracePeriod=600 Oct 04 03:14:02 crc kubenswrapper[4770]: I1004 03:14:02.531687 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="d700d3adbc22ca8f48ded121cdd0b6720fb293c613b3f9c60d656bef6938d1c8" exitCode=0 Oct 04 03:14:02 crc kubenswrapper[4770]: I1004 03:14:02.531743 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"d700d3adbc22ca8f48ded121cdd0b6720fb293c613b3f9c60d656bef6938d1c8"} Oct 04 03:14:02 crc kubenswrapper[4770]: I1004 03:14:02.532333 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"1f8ff4d68e47173d152e933c25298fdeacbf8df65acd72aab55c2b5b7e5c9c52"} Oct 04 03:14:02 crc kubenswrapper[4770]: I1004 03:14:02.532358 4770 scope.go:117] "RemoveContainer" containerID="52bc3793e50aecbadbef1d828c974de853af6a75a471a3fa3860ec13be9d7c91" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.165577 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-q2l8x"] Oct 04 03:14:16 crc kubenswrapper[4770]: E1004 03:14:16.166861 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="108c220a-bc76-4e17-884d-80f2a5b63357" containerName="registry" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.166896 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="108c220a-bc76-4e17-884d-80f2a5b63357" containerName="registry" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.167220 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="108c220a-bc76-4e17-884d-80f2a5b63357" containerName="registry" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.169230 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.174481 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.174571 4770 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-jz997" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.174736 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-q2l8x"] Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.175679 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.175900 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.306690 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-crc-storage\") pod \"crc-storage-crc-q2l8x\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.306789 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-node-mnt\") pod \"crc-storage-crc-q2l8x\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.307073 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whss7\" (UniqueName: \"kubernetes.io/projected/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-kube-api-access-whss7\") pod \"crc-storage-crc-q2l8x\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.408866 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-crc-storage\") pod \"crc-storage-crc-q2l8x\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.408975 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-node-mnt\") pod \"crc-storage-crc-q2l8x\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.409106 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whss7\" (UniqueName: \"kubernetes.io/projected/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-kube-api-access-whss7\") pod \"crc-storage-crc-q2l8x\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.409809 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-node-mnt\") pod \"crc-storage-crc-q2l8x\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.410781 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-crc-storage\") pod \"crc-storage-crc-q2l8x\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.449786 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whss7\" (UniqueName: \"kubernetes.io/projected/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-kube-api-access-whss7\") pod \"crc-storage-crc-q2l8x\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:16 crc kubenswrapper[4770]: I1004 03:14:16.535103 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:17 crc kubenswrapper[4770]: I1004 03:14:17.031738 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-q2l8x"] Oct 04 03:14:17 crc kubenswrapper[4770]: I1004 03:14:17.057959 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:14:17 crc kubenswrapper[4770]: I1004 03:14:17.644808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-q2l8x" event={"ID":"63e7d20a-bdd8-424e-a45f-85a2a405fcb7","Type":"ContainerStarted","Data":"eefae347c59dfa4bc54ef43e84a0eff37de896000644d16a7cbc239136c5b904"} Oct 04 03:14:18 crc kubenswrapper[4770]: I1004 03:14:18.654591 4770 generic.go:334] "Generic (PLEG): container finished" podID="63e7d20a-bdd8-424e-a45f-85a2a405fcb7" containerID="512ccd5465fdf014c659aa03b4fce7908acc47e9ff5722ef319b6a583f9c0c78" exitCode=0 Oct 04 03:14:18 crc kubenswrapper[4770]: I1004 03:14:18.654692 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-q2l8x" event={"ID":"63e7d20a-bdd8-424e-a45f-85a2a405fcb7","Type":"ContainerDied","Data":"512ccd5465fdf014c659aa03b4fce7908acc47e9ff5722ef319b6a583f9c0c78"} Oct 04 03:14:19 crc kubenswrapper[4770]: I1004 03:14:19.951410 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.070608 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whss7\" (UniqueName: \"kubernetes.io/projected/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-kube-api-access-whss7\") pod \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.071122 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-node-mnt\") pod \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.071167 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-crc-storage\") pod \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\" (UID: \"63e7d20a-bdd8-424e-a45f-85a2a405fcb7\") " Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.071317 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "63e7d20a-bdd8-424e-a45f-85a2a405fcb7" (UID: "63e7d20a-bdd8-424e-a45f-85a2a405fcb7"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.071931 4770 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.078309 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-kube-api-access-whss7" (OuterVolumeSpecName: "kube-api-access-whss7") pod "63e7d20a-bdd8-424e-a45f-85a2a405fcb7" (UID: "63e7d20a-bdd8-424e-a45f-85a2a405fcb7"). InnerVolumeSpecName "kube-api-access-whss7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.094349 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "63e7d20a-bdd8-424e-a45f-85a2a405fcb7" (UID: "63e7d20a-bdd8-424e-a45f-85a2a405fcb7"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.173568 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whss7\" (UniqueName: \"kubernetes.io/projected/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-kube-api-access-whss7\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.173652 4770 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/63e7d20a-bdd8-424e-a45f-85a2a405fcb7-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.669168 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-q2l8x" event={"ID":"63e7d20a-bdd8-424e-a45f-85a2a405fcb7","Type":"ContainerDied","Data":"eefae347c59dfa4bc54ef43e84a0eff37de896000644d16a7cbc239136c5b904"} Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.669213 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eefae347c59dfa4bc54ef43e84a0eff37de896000644d16a7cbc239136c5b904" Oct 04 03:14:20 crc kubenswrapper[4770]: I1004 03:14:20.669266 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-q2l8x" Oct 04 03:14:28 crc kubenswrapper[4770]: I1004 03:14:28.858390 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s"] Oct 04 03:14:28 crc kubenswrapper[4770]: E1004 03:14:28.859574 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e7d20a-bdd8-424e-a45f-85a2a405fcb7" containerName="storage" Oct 04 03:14:28 crc kubenswrapper[4770]: I1004 03:14:28.859603 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e7d20a-bdd8-424e-a45f-85a2a405fcb7" containerName="storage" Oct 04 03:14:28 crc kubenswrapper[4770]: I1004 03:14:28.859898 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="63e7d20a-bdd8-424e-a45f-85a2a405fcb7" containerName="storage" Oct 04 03:14:28 crc kubenswrapper[4770]: I1004 03:14:28.861304 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:28 crc kubenswrapper[4770]: I1004 03:14:28.872498 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 03:14:28 crc kubenswrapper[4770]: I1004 03:14:28.874307 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s"] Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.005610 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6p6x\" (UniqueName: \"kubernetes.io/projected/5e4248c4-b986-498f-b700-f56c4eb3e29d-kube-api-access-n6p6x\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.005915 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.006079 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.107857 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.107984 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6p6x\" (UniqueName: \"kubernetes.io/projected/5e4248c4-b986-498f-b700-f56c4eb3e29d-kube-api-access-n6p6x\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.108226 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.109236 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.109428 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.136682 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6p6x\" (UniqueName: \"kubernetes.io/projected/5e4248c4-b986-498f-b700-f56c4eb3e29d-kube-api-access-n6p6x\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.191701 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.454504 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s"] Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.731895 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" event={"ID":"5e4248c4-b986-498f-b700-f56c4eb3e29d","Type":"ContainerStarted","Data":"261be52aa1b0fb900aeb2839d4fe94c442e3d43ccf4ee6d7ecb9177f9e6462e1"} Oct 04 03:14:29 crc kubenswrapper[4770]: I1004 03:14:29.732061 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" event={"ID":"5e4248c4-b986-498f-b700-f56c4eb3e29d","Type":"ContainerStarted","Data":"c78e50e77179110da15ba21f14bd44206595875f1e08ac70190cc0119f27890c"} Oct 04 03:14:30 crc kubenswrapper[4770]: I1004 03:14:30.740196 4770 generic.go:334] "Generic (PLEG): container finished" podID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerID="261be52aa1b0fb900aeb2839d4fe94c442e3d43ccf4ee6d7ecb9177f9e6462e1" exitCode=0 Oct 04 03:14:30 crc kubenswrapper[4770]: I1004 03:14:30.740277 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" event={"ID":"5e4248c4-b986-498f-b700-f56c4eb3e29d","Type":"ContainerDied","Data":"261be52aa1b0fb900aeb2839d4fe94c442e3d43ccf4ee6d7ecb9177f9e6462e1"} Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.528062 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rtksw"] Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.528755 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovn-controller" containerID="cri-o://cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514" gracePeriod=30 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.529079 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="sbdb" containerID="cri-o://fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea" gracePeriod=30 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.529130 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kube-rbac-proxy-node" containerID="cri-o://56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b" gracePeriod=30 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.529164 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovn-acl-logging" containerID="cri-o://2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d" gracePeriod=30 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.529163 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="nbdb" containerID="cri-o://d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1" gracePeriod=30 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.529179 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e" gracePeriod=30 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.529226 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="northd" containerID="cri-o://24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce" gracePeriod=30 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.578861 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" containerID="cri-o://80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50" gracePeriod=30 Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.597694 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.598421 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.600689 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.601045 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.608172 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.608265 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="sbdb" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.611132 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.611168 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="nbdb" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.754157 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovnkube-controller/3.log" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.759648 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovn-acl-logging/0.log" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760202 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovn-controller/0.log" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760569 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50" exitCode=0 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760601 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1" exitCode=0 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760611 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e" exitCode=0 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760621 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b" exitCode=0 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760631 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d" exitCode=143 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760654 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514" exitCode=143 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760715 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50"} Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760766 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1"} Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760778 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e"} Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760788 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b"} Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760798 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d"} Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514"} Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.760826 4770 scope.go:117] "RemoveContainer" containerID="fcd236a1b2c16f3c1895d2a27d49bf46fb309f58c6e4231fba21d1009773c653" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.763874 4770 generic.go:334] "Generic (PLEG): container finished" podID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerID="3dfab23de50ef0be5d4c34428c26fe89ec82fcd3a26ff61019e7d48b030aea40" exitCode=0 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.763997 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" event={"ID":"5e4248c4-b986-498f-b700-f56c4eb3e29d","Type":"ContainerDied","Data":"3dfab23de50ef0be5d4c34428c26fe89ec82fcd3a26ff61019e7d48b030aea40"} Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.766816 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/2.log" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.767239 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/1.log" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.767277 4770 generic.go:334] "Generic (PLEG): container finished" podID="88d62119-65f3-4846-8b93-53d096fa9df7" containerID="9f37453b28a502cbdaf06be847a0a05df4f86f3ee80d189151769d8da5ab3c91" exitCode=2 Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.767299 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-66tn4" event={"ID":"88d62119-65f3-4846-8b93-53d096fa9df7","Type":"ContainerDied","Data":"9f37453b28a502cbdaf06be847a0a05df4f86f3ee80d189151769d8da5ab3c91"} Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.768423 4770 scope.go:117] "RemoveContainer" containerID="9f37453b28a502cbdaf06be847a0a05df4f86f3ee80d189151769d8da5ab3c91" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.768624 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-66tn4_openshift-multus(88d62119-65f3-4846-8b93-53d096fa9df7)\"" pod="openshift-multus/multus-66tn4" podUID="88d62119-65f3-4846-8b93-53d096fa9df7" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.851178 4770 scope.go:117] "RemoveContainer" containerID="6cd9a0ab71a23ca260835682f5020a316049fdb2b52eb013a0154b7d9371cad0" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.886346 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovn-acl-logging/0.log" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.889053 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovn-controller/0.log" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.889977 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955351 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gdbk9"] Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955559 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955571 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955579 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="northd" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955585 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="northd" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955593 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955600 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955606 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955612 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955619 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="nbdb" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955625 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="nbdb" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955633 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovn-acl-logging" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955639 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovn-acl-logging" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955646 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kubecfg-setup" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955652 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kubecfg-setup" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955659 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="sbdb" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955664 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="sbdb" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955672 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955678 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955684 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kube-rbac-proxy-node" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955690 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kube-rbac-proxy-node" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955700 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955706 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.955715 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovn-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955720 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovn-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955824 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovn-acl-logging" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955834 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955843 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955849 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955855 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955863 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="northd" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955871 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="sbdb" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955880 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovn-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955886 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="nbdb" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.955894 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="kube-rbac-proxy-node" Oct 04 03:14:32 crc kubenswrapper[4770]: E1004 03:14:32.956000 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.956023 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.956129 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.956138 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerName="ovnkube-controller" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.957692 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.957946 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-systemd\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958021 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovn-node-metrics-cert\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958054 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-config\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958069 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-kubelet\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958103 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-openvswitch\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958119 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-script-lib\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958136 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg9hp\" (UniqueName: \"kubernetes.io/projected/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-kube-api-access-dg9hp\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958159 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958174 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-netns\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958189 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-slash\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958214 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-env-overrides\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958229 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-log-socket\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958251 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-etc-openvswitch\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958264 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-var-lib-openvswitch\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958281 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-netd\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958295 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-node-log\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958310 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-ovn-kubernetes\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958351 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-ovn\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958388 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-systemd-units\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958403 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-bin\") pod \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\" (UID: \"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2\") " Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.958628 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.962390 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-slash" (OuterVolumeSpecName: "host-slash") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963112 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963418 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963531 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-log-socket" (OuterVolumeSpecName: "log-socket") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963477 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963503 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963530 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963570 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963597 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963623 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963649 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963662 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963674 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963700 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.963880 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.964313 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-node-log" (OuterVolumeSpecName: "node-log") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.972424 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.977647 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:32 crc kubenswrapper[4770]: I1004 03:14:32.979335 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-kube-api-access-dg9hp" (OuterVolumeSpecName: "kube-api-access-dg9hp") pod "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" (UID: "61fd243e-61c3-4f86-8ee3-5c374a3e8ce2"). InnerVolumeSpecName "kube-api-access-dg9hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.059942 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-node-log\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060000 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060042 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-cni-bin\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060058 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-cni-netd\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060077 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-var-lib-openvswitch\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060102 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-run-ovn\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060116 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-slash\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060130 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldwrx\" (UniqueName: \"kubernetes.io/projected/4053a266-fc27-475b-8a78-99f3fcec60be-kube-api-access-ldwrx\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060144 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-run-openvswitch\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060157 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-log-socket\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060173 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-kubelet\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060192 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-systemd-units\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060211 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4053a266-fc27-475b-8a78-99f3fcec60be-env-overrides\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060227 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-run-systemd\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060263 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-etc-openvswitch\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060284 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-run-ovn-kubernetes\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060308 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4053a266-fc27-475b-8a78-99f3fcec60be-ovnkube-config\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060325 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-run-netns\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060347 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4053a266-fc27-475b-8a78-99f3fcec60be-ovnkube-script-lib\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060367 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4053a266-fc27-475b-8a78-99f3fcec60be-ovn-node-metrics-cert\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060420 4770 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060431 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060444 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060455 4770 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060464 4770 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060473 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg9hp\" (UniqueName: \"kubernetes.io/projected/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-kube-api-access-dg9hp\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060482 4770 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060492 4770 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060501 4770 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060510 4770 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-slash\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060518 4770 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060526 4770 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-log-socket\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060535 4770 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060544 4770 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060552 4770 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060560 4770 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060568 4770 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-node-log\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060577 4770 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060584 4770 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.060593 4770 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.161815 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-run-ovn-kubernetes\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.161872 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4053a266-fc27-475b-8a78-99f3fcec60be-ovnkube-config\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.161905 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-run-netns\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.161922 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4053a266-fc27-475b-8a78-99f3fcec60be-ovnkube-script-lib\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.161943 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4053a266-fc27-475b-8a78-99f3fcec60be-ovn-node-metrics-cert\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.161965 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-node-log\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.161969 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-run-ovn-kubernetes\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162063 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.161987 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162114 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-cni-bin\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162132 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-cni-netd\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162148 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-var-lib-openvswitch\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162171 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-run-ovn\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162184 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-slash\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162200 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldwrx\" (UniqueName: \"kubernetes.io/projected/4053a266-fc27-475b-8a78-99f3fcec60be-kube-api-access-ldwrx\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162218 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-run-openvswitch\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162233 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-log-socket\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162252 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-kubelet\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162273 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-systemd-units\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162295 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4053a266-fc27-475b-8a78-99f3fcec60be-env-overrides\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162310 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-run-systemd\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162353 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-etc-openvswitch\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162371 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-var-lib-openvswitch\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162426 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-run-ovn\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162459 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-slash\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162747 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-run-openvswitch\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162775 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-log-socket\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162795 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-kubelet\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162820 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-systemd-units\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162886 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4053a266-fc27-475b-8a78-99f3fcec60be-ovnkube-script-lib\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162895 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4053a266-fc27-475b-8a78-99f3fcec60be-ovnkube-config\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162927 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-run-netns\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162950 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-run-systemd\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162960 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-cni-bin\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162987 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-host-cni-netd\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162988 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-node-log\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.162399 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4053a266-fc27-475b-8a78-99f3fcec60be-etc-openvswitch\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.163171 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4053a266-fc27-475b-8a78-99f3fcec60be-env-overrides\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.168312 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4053a266-fc27-475b-8a78-99f3fcec60be-ovn-node-metrics-cert\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.177869 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldwrx\" (UniqueName: \"kubernetes.io/projected/4053a266-fc27-475b-8a78-99f3fcec60be-kube-api-access-ldwrx\") pod \"ovnkube-node-gdbk9\" (UID: \"4053a266-fc27-475b-8a78-99f3fcec60be\") " pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.287601 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:33 crc kubenswrapper[4770]: W1004 03:14:33.316002 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4053a266_fc27_475b_8a78_99f3fcec60be.slice/crio-8b3e07e86b4a8b70172ec7b57dea33848f05aced926cd012305a20ff2518c8c7 WatchSource:0}: Error finding container 8b3e07e86b4a8b70172ec7b57dea33848f05aced926cd012305a20ff2518c8c7: Status 404 returned error can't find the container with id 8b3e07e86b4a8b70172ec7b57dea33848f05aced926cd012305a20ff2518c8c7 Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.772274 4770 generic.go:334] "Generic (PLEG): container finished" podID="4053a266-fc27-475b-8a78-99f3fcec60be" containerID="2251c107399a2e02bf68e4910f59f712107f9e99f16a46434aede45760e959f9" exitCode=0 Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.772356 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerDied","Data":"2251c107399a2e02bf68e4910f59f712107f9e99f16a46434aede45760e959f9"} Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.772391 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerStarted","Data":"8b3e07e86b4a8b70172ec7b57dea33848f05aced926cd012305a20ff2518c8c7"} Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.775798 4770 generic.go:334] "Generic (PLEG): container finished" podID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerID="ee0d1afb37fb4dada33f8613fd486409b865a78db552f29e52aa8f16f9ba07e5" exitCode=0 Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.775857 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" event={"ID":"5e4248c4-b986-498f-b700-f56c4eb3e29d","Type":"ContainerDied","Data":"ee0d1afb37fb4dada33f8613fd486409b865a78db552f29e52aa8f16f9ba07e5"} Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.785929 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovn-acl-logging/0.log" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.786823 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rtksw_61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/ovn-controller/0.log" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.787351 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea" exitCode=0 Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.787376 4770 generic.go:334] "Generic (PLEG): container finished" podID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" containerID="24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce" exitCode=0 Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.787436 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea"} Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.787463 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce"} Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.787478 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" event={"ID":"61fd243e-61c3-4f86-8ee3-5c374a3e8ce2","Type":"ContainerDied","Data":"2fd94e16a6226dc24440e4f8e4db5b9cbbd86e0323ba12c7fa7fcbe8c884900b"} Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.787461 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rtksw" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.787572 4770 scope.go:117] "RemoveContainer" containerID="80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.790216 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/2.log" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.839049 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rtksw"] Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.843515 4770 scope.go:117] "RemoveContainer" containerID="fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.846377 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rtksw"] Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.871670 4770 scope.go:117] "RemoveContainer" containerID="d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.895819 4770 scope.go:117] "RemoveContainer" containerID="24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.915920 4770 scope.go:117] "RemoveContainer" containerID="91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.949407 4770 scope.go:117] "RemoveContainer" containerID="56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.979708 4770 scope.go:117] "RemoveContainer" containerID="2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d" Oct 04 03:14:33 crc kubenswrapper[4770]: I1004 03:14:33.994045 4770 scope.go:117] "RemoveContainer" containerID="cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.026037 4770 scope.go:117] "RemoveContainer" containerID="95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.048919 4770 scope.go:117] "RemoveContainer" containerID="80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50" Oct 04 03:14:34 crc kubenswrapper[4770]: E1004 03:14:34.049442 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50\": container with ID starting with 80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50 not found: ID does not exist" containerID="80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.049493 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50"} err="failed to get container status \"80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50\": rpc error: code = NotFound desc = could not find container \"80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50\": container with ID starting with 80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50 not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.049526 4770 scope.go:117] "RemoveContainer" containerID="fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea" Oct 04 03:14:34 crc kubenswrapper[4770]: E1004 03:14:34.049849 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\": container with ID starting with fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea not found: ID does not exist" containerID="fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.049876 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea"} err="failed to get container status \"fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\": rpc error: code = NotFound desc = could not find container \"fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\": container with ID starting with fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.049894 4770 scope.go:117] "RemoveContainer" containerID="d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1" Oct 04 03:14:34 crc kubenswrapper[4770]: E1004 03:14:34.050156 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\": container with ID starting with d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1 not found: ID does not exist" containerID="d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.050185 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1"} err="failed to get container status \"d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\": rpc error: code = NotFound desc = could not find container \"d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\": container with ID starting with d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1 not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.050203 4770 scope.go:117] "RemoveContainer" containerID="24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce" Oct 04 03:14:34 crc kubenswrapper[4770]: E1004 03:14:34.050417 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\": container with ID starting with 24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce not found: ID does not exist" containerID="24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.050446 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce"} err="failed to get container status \"24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\": rpc error: code = NotFound desc = could not find container \"24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\": container with ID starting with 24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.050466 4770 scope.go:117] "RemoveContainer" containerID="91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e" Oct 04 03:14:34 crc kubenswrapper[4770]: E1004 03:14:34.050680 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\": container with ID starting with 91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e not found: ID does not exist" containerID="91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.050705 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e"} err="failed to get container status \"91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\": rpc error: code = NotFound desc = could not find container \"91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\": container with ID starting with 91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.050722 4770 scope.go:117] "RemoveContainer" containerID="56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b" Oct 04 03:14:34 crc kubenswrapper[4770]: E1004 03:14:34.050940 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\": container with ID starting with 56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b not found: ID does not exist" containerID="56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.050963 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b"} err="failed to get container status \"56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\": rpc error: code = NotFound desc = could not find container \"56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\": container with ID starting with 56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.051000 4770 scope.go:117] "RemoveContainer" containerID="2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d" Oct 04 03:14:34 crc kubenswrapper[4770]: E1004 03:14:34.051255 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\": container with ID starting with 2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d not found: ID does not exist" containerID="2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.051284 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d"} err="failed to get container status \"2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\": rpc error: code = NotFound desc = could not find container \"2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\": container with ID starting with 2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.051300 4770 scope.go:117] "RemoveContainer" containerID="cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514" Oct 04 03:14:34 crc kubenswrapper[4770]: E1004 03:14:34.051830 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\": container with ID starting with cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514 not found: ID does not exist" containerID="cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.051857 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514"} err="failed to get container status \"cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\": rpc error: code = NotFound desc = could not find container \"cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\": container with ID starting with cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514 not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.051873 4770 scope.go:117] "RemoveContainer" containerID="95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a" Oct 04 03:14:34 crc kubenswrapper[4770]: E1004 03:14:34.052163 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\": container with ID starting with 95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a not found: ID does not exist" containerID="95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.052189 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a"} err="failed to get container status \"95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\": rpc error: code = NotFound desc = could not find container \"95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\": container with ID starting with 95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.052205 4770 scope.go:117] "RemoveContainer" containerID="80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.052473 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50"} err="failed to get container status \"80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50\": rpc error: code = NotFound desc = could not find container \"80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50\": container with ID starting with 80ca835fb8d11e9c1c9ea780f27946029afffe11b4a29eee5546b7e80f584f50 not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.052497 4770 scope.go:117] "RemoveContainer" containerID="fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.052866 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea"} err="failed to get container status \"fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\": rpc error: code = NotFound desc = could not find container \"fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea\": container with ID starting with fd95e6ccabe6b3472deece060ae758b4221db7f59c2846d5049c642b83a24fea not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.052894 4770 scope.go:117] "RemoveContainer" containerID="d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.053242 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1"} err="failed to get container status \"d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\": rpc error: code = NotFound desc = could not find container \"d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1\": container with ID starting with d24f2ca579514f73b11cb61224bbe7a6dc50018cd3b708f4b7533c66d0d4ddf1 not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.053316 4770 scope.go:117] "RemoveContainer" containerID="24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.053653 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce"} err="failed to get container status \"24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\": rpc error: code = NotFound desc = could not find container \"24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce\": container with ID starting with 24032b3212afc662043664b08ad5afebfa0c54207e170cdd66cf156c120380ce not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.053684 4770 scope.go:117] "RemoveContainer" containerID="91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.053968 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e"} err="failed to get container status \"91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\": rpc error: code = NotFound desc = could not find container \"91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e\": container with ID starting with 91f773f743ae35aac9952b5b953b7898e9686d57a7797ef5d8e71dacedd2c39e not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.053993 4770 scope.go:117] "RemoveContainer" containerID="56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.054413 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b"} err="failed to get container status \"56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\": rpc error: code = NotFound desc = could not find container \"56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b\": container with ID starting with 56170ad6c502ed6a2d814452cee39e744d4865da562d312be7cb0fdbc5fd649b not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.054459 4770 scope.go:117] "RemoveContainer" containerID="2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.054830 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d"} err="failed to get container status \"2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\": rpc error: code = NotFound desc = could not find container \"2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d\": container with ID starting with 2cea18dff47745080c8de7c7479128e7a8ee3c3d4cfe73dc7e2107cffb0bef2d not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.054865 4770 scope.go:117] "RemoveContainer" containerID="cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.055253 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514"} err="failed to get container status \"cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\": rpc error: code = NotFound desc = could not find container \"cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514\": container with ID starting with cbb1075c93f5401f7f35d62409fc7ebe4937b8ccdac51b9e03589d8f38f0a514 not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.055283 4770 scope.go:117] "RemoveContainer" containerID="95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.055616 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a"} err="failed to get container status \"95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\": rpc error: code = NotFound desc = could not find container \"95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a\": container with ID starting with 95decd673ce3917c57fd72460eab937ede259425e3fd44dc58de0f9fbcf68e7a not found: ID does not exist" Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.808701 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerStarted","Data":"3143a3953972a106350cb4318d96ea3fee70e25194da2697ca514e9fbfb6324c"} Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.809272 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerStarted","Data":"1b469e2fe82264ffedf29609e561dfd6770f52a7773c10977490985954b993df"} Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.809301 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerStarted","Data":"9157f7870d6d3822bcee43909168429e2e8aee03fb7ae71a6f4f23b98a1f9cd8"} Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.809323 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerStarted","Data":"f917d61e892af3efd133b75a755142575eee69622e7b33409aa2824f7cf70e11"} Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.809345 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerStarted","Data":"b84411ba48280d721976aeb723dfaf79a6e7cb3a9e8fdd84d66c69fc6ddbb5f1"} Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.809370 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerStarted","Data":"6e292c5a7f6a6852616358cb3c1e8b5a2fe44afb9e0437cfbfa0b23a9fe6862d"} Oct 04 03:14:34 crc kubenswrapper[4770]: I1004 03:14:34.917765 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.011352 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6p6x\" (UniqueName: \"kubernetes.io/projected/5e4248c4-b986-498f-b700-f56c4eb3e29d-kube-api-access-n6p6x\") pod \"5e4248c4-b986-498f-b700-f56c4eb3e29d\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.011482 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-util\") pod \"5e4248c4-b986-498f-b700-f56c4eb3e29d\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.011520 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-bundle\") pod \"5e4248c4-b986-498f-b700-f56c4eb3e29d\" (UID: \"5e4248c4-b986-498f-b700-f56c4eb3e29d\") " Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.012619 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-bundle" (OuterVolumeSpecName: "bundle") pod "5e4248c4-b986-498f-b700-f56c4eb3e29d" (UID: "5e4248c4-b986-498f-b700-f56c4eb3e29d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.022556 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e4248c4-b986-498f-b700-f56c4eb3e29d-kube-api-access-n6p6x" (OuterVolumeSpecName: "kube-api-access-n6p6x") pod "5e4248c4-b986-498f-b700-f56c4eb3e29d" (UID: "5e4248c4-b986-498f-b700-f56c4eb3e29d"). InnerVolumeSpecName "kube-api-access-n6p6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.041568 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-util" (OuterVolumeSpecName: "util") pod "5e4248c4-b986-498f-b700-f56c4eb3e29d" (UID: "5e4248c4-b986-498f-b700-f56c4eb3e29d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.113312 4770 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.113362 4770 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e4248c4-b986-498f-b700-f56c4eb3e29d-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.113383 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6p6x\" (UniqueName: \"kubernetes.io/projected/5e4248c4-b986-498f-b700-f56c4eb3e29d-kube-api-access-n6p6x\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.685774 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61fd243e-61c3-4f86-8ee3-5c374a3e8ce2" path="/var/lib/kubelet/pods/61fd243e-61c3-4f86-8ee3-5c374a3e8ce2/volumes" Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.821161 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" event={"ID":"5e4248c4-b986-498f-b700-f56c4eb3e29d","Type":"ContainerDied","Data":"c78e50e77179110da15ba21f14bd44206595875f1e08ac70190cc0119f27890c"} Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.821209 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c78e50e77179110da15ba21f14bd44206595875f1e08ac70190cc0119f27890c" Oct 04 03:14:35 crc kubenswrapper[4770]: I1004 03:14:35.821224 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s" Oct 04 03:14:37 crc kubenswrapper[4770]: I1004 03:14:37.838346 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerStarted","Data":"11e266167ea501dfbc5836c0b921b0942945998f8f6c6e1d247e59456300d029"} Oct 04 03:14:39 crc kubenswrapper[4770]: I1004 03:14:39.855155 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" event={"ID":"4053a266-fc27-475b-8a78-99f3fcec60be","Type":"ContainerStarted","Data":"f6c8aa5a074a5630f94f143c8c3b706d33fed55f244eca44134a5588bcbf3724"} Oct 04 03:14:39 crc kubenswrapper[4770]: I1004 03:14:39.855951 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:39 crc kubenswrapper[4770]: I1004 03:14:39.855975 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:39 crc kubenswrapper[4770]: I1004 03:14:39.887847 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:39 crc kubenswrapper[4770]: I1004 03:14:39.890125 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" podStartSLOduration=7.890094511 podStartE2EDuration="7.890094511s" podCreationTimestamp="2025-10-04 03:14:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:14:39.884513336 +0000 UTC m=+691.176523058" watchObservedRunningTime="2025-10-04 03:14:39.890094511 +0000 UTC m=+691.182104253" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.254081 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-94wmz"] Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.254295 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerName="util" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.254341 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerName="util" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.254354 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerName="extract" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.254360 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerName="extract" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.254378 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerName="pull" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.254386 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerName="pull" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.254468 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e4248c4-b986-498f-b700-f56c4eb3e29d" containerName="extract" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.254829 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.259816 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.260038 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.260595 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-lvc54" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.288483 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-94wmz"] Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.388666 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85gmn\" (UniqueName: \"kubernetes.io/projected/2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb-kube-api-access-85gmn\") pod \"nmstate-operator-858ddd8f98-94wmz\" (UID: \"2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.490622 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85gmn\" (UniqueName: \"kubernetes.io/projected/2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb-kube-api-access-85gmn\") pod \"nmstate-operator-858ddd8f98-94wmz\" (UID: \"2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.516871 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85gmn\" (UniqueName: \"kubernetes.io/projected/2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb-kube-api-access-85gmn\") pod \"nmstate-operator-858ddd8f98-94wmz\" (UID: \"2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.586957 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.630843 4770 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(c91d22c4410cb405a3cd215b9b62c7d39e51b0e50e9030d5e7ac784560e2c619): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.630942 4770 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(c91d22c4410cb405a3cd215b9b62c7d39e51b0e50e9030d5e7ac784560e2c619): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.630967 4770 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(c91d22c4410cb405a3cd215b9b62c7d39e51b0e50e9030d5e7ac784560e2c619): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.631043 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nmstate-operator-858ddd8f98-94wmz_openshift-nmstate(2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nmstate-operator-858ddd8f98-94wmz_openshift-nmstate(2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(c91d22c4410cb405a3cd215b9b62c7d39e51b0e50e9030d5e7ac784560e2c619): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" podUID="2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.863791 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.864673 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.865084 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.909446 4770 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(720683bd9560566c2a7bc4f51e63b19d4e24a1ff7fd7ba493b341055a70f64ca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.909577 4770 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(720683bd9560566c2a7bc4f51e63b19d4e24a1ff7fd7ba493b341055a70f64ca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.909626 4770 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(720683bd9560566c2a7bc4f51e63b19d4e24a1ff7fd7ba493b341055a70f64ca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:40 crc kubenswrapper[4770]: E1004 03:14:40.909728 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nmstate-operator-858ddd8f98-94wmz_openshift-nmstate(2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nmstate-operator-858ddd8f98-94wmz_openshift-nmstate(2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(720683bd9560566c2a7bc4f51e63b19d4e24a1ff7fd7ba493b341055a70f64ca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" podUID="2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb" Oct 04 03:14:40 crc kubenswrapper[4770]: I1004 03:14:40.916236 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:14:44 crc kubenswrapper[4770]: I1004 03:14:44.673862 4770 scope.go:117] "RemoveContainer" containerID="9f37453b28a502cbdaf06be847a0a05df4f86f3ee80d189151769d8da5ab3c91" Oct 04 03:14:44 crc kubenswrapper[4770]: E1004 03:14:44.674576 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-66tn4_openshift-multus(88d62119-65f3-4846-8b93-53d096fa9df7)\"" pod="openshift-multus/multus-66tn4" podUID="88d62119-65f3-4846-8b93-53d096fa9df7" Oct 04 03:14:52 crc kubenswrapper[4770]: I1004 03:14:52.673174 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:52 crc kubenswrapper[4770]: I1004 03:14:52.674658 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:52 crc kubenswrapper[4770]: E1004 03:14:52.716455 4770 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(8c1588e52b7fab35ee535d4e4b04a3abb445a2c6d7bc9d00fb170b1d01e6e5ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:14:52 crc kubenswrapper[4770]: E1004 03:14:52.716571 4770 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(8c1588e52b7fab35ee535d4e4b04a3abb445a2c6d7bc9d00fb170b1d01e6e5ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:52 crc kubenswrapper[4770]: E1004 03:14:52.716606 4770 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(8c1588e52b7fab35ee535d4e4b04a3abb445a2c6d7bc9d00fb170b1d01e6e5ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:14:52 crc kubenswrapper[4770]: E1004 03:14:52.716705 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nmstate-operator-858ddd8f98-94wmz_openshift-nmstate(2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nmstate-operator-858ddd8f98-94wmz_openshift-nmstate(2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-94wmz_openshift-nmstate_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb_0(8c1588e52b7fab35ee535d4e4b04a3abb445a2c6d7bc9d00fb170b1d01e6e5ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" podUID="2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb" Oct 04 03:14:58 crc kubenswrapper[4770]: I1004 03:14:58.674355 4770 scope.go:117] "RemoveContainer" containerID="9f37453b28a502cbdaf06be847a0a05df4f86f3ee80d189151769d8da5ab3c91" Oct 04 03:14:58 crc kubenswrapper[4770]: I1004 03:14:58.987271 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-66tn4_88d62119-65f3-4846-8b93-53d096fa9df7/kube-multus/2.log" Oct 04 03:14:58 crc kubenswrapper[4770]: I1004 03:14:58.987672 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-66tn4" event={"ID":"88d62119-65f3-4846-8b93-53d096fa9df7","Type":"ContainerStarted","Data":"a31b11bdc74d6ef358eb2895a30a4cf98e4df581a2ba585613b30f8d731999f9"} Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.148035 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w"] Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.149611 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.154463 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.154713 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.159836 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w"] Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.244111 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6739a0a5-d8da-49e3-8131-d2e9607a0455-secret-volume\") pod \"collect-profiles-29325795-hrw4w\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.244758 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gckcp\" (UniqueName: \"kubernetes.io/projected/6739a0a5-d8da-49e3-8131-d2e9607a0455-kube-api-access-gckcp\") pod \"collect-profiles-29325795-hrw4w\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.244966 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6739a0a5-d8da-49e3-8131-d2e9607a0455-config-volume\") pod \"collect-profiles-29325795-hrw4w\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.348171 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6739a0a5-d8da-49e3-8131-d2e9607a0455-secret-volume\") pod \"collect-profiles-29325795-hrw4w\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.348257 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gckcp\" (UniqueName: \"kubernetes.io/projected/6739a0a5-d8da-49e3-8131-d2e9607a0455-kube-api-access-gckcp\") pod \"collect-profiles-29325795-hrw4w\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.348288 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6739a0a5-d8da-49e3-8131-d2e9607a0455-config-volume\") pod \"collect-profiles-29325795-hrw4w\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.350350 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6739a0a5-d8da-49e3-8131-d2e9607a0455-config-volume\") pod \"collect-profiles-29325795-hrw4w\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.363412 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6739a0a5-d8da-49e3-8131-d2e9607a0455-secret-volume\") pod \"collect-profiles-29325795-hrw4w\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.374806 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gckcp\" (UniqueName: \"kubernetes.io/projected/6739a0a5-d8da-49e3-8131-d2e9607a0455-kube-api-access-gckcp\") pod \"collect-profiles-29325795-hrw4w\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.474743 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:00 crc kubenswrapper[4770]: I1004 03:15:00.767891 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w"] Oct 04 03:15:00 crc kubenswrapper[4770]: W1004 03:15:00.778643 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6739a0a5_d8da_49e3_8131_d2e9607a0455.slice/crio-6a95cc709e6d2f1aa6a6dc8c257d15223490786dd148189741e346f9cb36ac3d WatchSource:0}: Error finding container 6a95cc709e6d2f1aa6a6dc8c257d15223490786dd148189741e346f9cb36ac3d: Status 404 returned error can't find the container with id 6a95cc709e6d2f1aa6a6dc8c257d15223490786dd148189741e346f9cb36ac3d Oct 04 03:15:01 crc kubenswrapper[4770]: I1004 03:15:01.006121 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" event={"ID":"6739a0a5-d8da-49e3-8131-d2e9607a0455","Type":"ContainerStarted","Data":"4bd7c972b5ebb9e930952a683340a6cad67fb4154f04dd3135db420e4dc96a73"} Oct 04 03:15:01 crc kubenswrapper[4770]: I1004 03:15:01.006180 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" event={"ID":"6739a0a5-d8da-49e3-8131-d2e9607a0455","Type":"ContainerStarted","Data":"6a95cc709e6d2f1aa6a6dc8c257d15223490786dd148189741e346f9cb36ac3d"} Oct 04 03:15:01 crc kubenswrapper[4770]: I1004 03:15:01.028583 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" podStartSLOduration=1.028564473 podStartE2EDuration="1.028564473s" podCreationTimestamp="2025-10-04 03:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:15:01.025878363 +0000 UTC m=+712.317888145" watchObservedRunningTime="2025-10-04 03:15:01.028564473 +0000 UTC m=+712.320574185" Oct 04 03:15:02 crc kubenswrapper[4770]: I1004 03:15:02.014936 4770 generic.go:334] "Generic (PLEG): container finished" podID="6739a0a5-d8da-49e3-8131-d2e9607a0455" containerID="4bd7c972b5ebb9e930952a683340a6cad67fb4154f04dd3135db420e4dc96a73" exitCode=0 Oct 04 03:15:02 crc kubenswrapper[4770]: I1004 03:15:02.015090 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" event={"ID":"6739a0a5-d8da-49e3-8131-d2e9607a0455","Type":"ContainerDied","Data":"4bd7c972b5ebb9e930952a683340a6cad67fb4154f04dd3135db420e4dc96a73"} Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.325821 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gdbk9" Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.337568 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.396659 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6739a0a5-d8da-49e3-8131-d2e9607a0455-secret-volume\") pod \"6739a0a5-d8da-49e3-8131-d2e9607a0455\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.396738 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gckcp\" (UniqueName: \"kubernetes.io/projected/6739a0a5-d8da-49e3-8131-d2e9607a0455-kube-api-access-gckcp\") pod \"6739a0a5-d8da-49e3-8131-d2e9607a0455\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.396774 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6739a0a5-d8da-49e3-8131-d2e9607a0455-config-volume\") pod \"6739a0a5-d8da-49e3-8131-d2e9607a0455\" (UID: \"6739a0a5-d8da-49e3-8131-d2e9607a0455\") " Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.398310 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6739a0a5-d8da-49e3-8131-d2e9607a0455-config-volume" (OuterVolumeSpecName: "config-volume") pod "6739a0a5-d8da-49e3-8131-d2e9607a0455" (UID: "6739a0a5-d8da-49e3-8131-d2e9607a0455"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.406959 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6739a0a5-d8da-49e3-8131-d2e9607a0455-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6739a0a5-d8da-49e3-8131-d2e9607a0455" (UID: "6739a0a5-d8da-49e3-8131-d2e9607a0455"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.407045 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6739a0a5-d8da-49e3-8131-d2e9607a0455-kube-api-access-gckcp" (OuterVolumeSpecName: "kube-api-access-gckcp") pod "6739a0a5-d8da-49e3-8131-d2e9607a0455" (UID: "6739a0a5-d8da-49e3-8131-d2e9607a0455"). InnerVolumeSpecName "kube-api-access-gckcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.498795 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6739a0a5-d8da-49e3-8131-d2e9607a0455-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.498843 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gckcp\" (UniqueName: \"kubernetes.io/projected/6739a0a5-d8da-49e3-8131-d2e9607a0455-kube-api-access-gckcp\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:03 crc kubenswrapper[4770]: I1004 03:15:03.498862 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6739a0a5-d8da-49e3-8131-d2e9607a0455-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:04 crc kubenswrapper[4770]: I1004 03:15:04.031188 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" event={"ID":"6739a0a5-d8da-49e3-8131-d2e9607a0455","Type":"ContainerDied","Data":"6a95cc709e6d2f1aa6a6dc8c257d15223490786dd148189741e346f9cb36ac3d"} Oct 04 03:15:04 crc kubenswrapper[4770]: I1004 03:15:04.031271 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a95cc709e6d2f1aa6a6dc8c257d15223490786dd148189741e346f9cb36ac3d" Oct 04 03:15:04 crc kubenswrapper[4770]: I1004 03:15:04.031327 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w" Oct 04 03:15:06 crc kubenswrapper[4770]: I1004 03:15:06.673088 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:15:06 crc kubenswrapper[4770]: I1004 03:15:06.673923 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" Oct 04 03:15:06 crc kubenswrapper[4770]: I1004 03:15:06.905925 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-94wmz"] Oct 04 03:15:07 crc kubenswrapper[4770]: I1004 03:15:07.056503 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" event={"ID":"2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb","Type":"ContainerStarted","Data":"6b1c755d1aadcc43c27ceb36403e290493324375ca01202ba26e075d79567530"} Oct 04 03:15:10 crc kubenswrapper[4770]: I1004 03:15:10.083514 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" event={"ID":"2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb","Type":"ContainerStarted","Data":"7314bae942865d198cb1faffe1cacc3939bc0a074fc9b915b597ea4314d90a61"} Oct 04 03:15:10 crc kubenswrapper[4770]: I1004 03:15:10.108501 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-94wmz" podStartSLOduration=27.608119186 podStartE2EDuration="30.108474179s" podCreationTimestamp="2025-10-04 03:14:40 +0000 UTC" firstStartedPulling="2025-10-04 03:15:06.915726368 +0000 UTC m=+718.207736080" lastFinishedPulling="2025-10-04 03:15:09.416081321 +0000 UTC m=+720.708091073" observedRunningTime="2025-10-04 03:15:10.105673586 +0000 UTC m=+721.397683298" watchObservedRunningTime="2025-10-04 03:15:10.108474179 +0000 UTC m=+721.400483901" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.183863 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j"] Oct 04 03:15:11 crc kubenswrapper[4770]: E1004 03:15:11.184607 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6739a0a5-d8da-49e3-8131-d2e9607a0455" containerName="collect-profiles" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.184633 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6739a0a5-d8da-49e3-8131-d2e9607a0455" containerName="collect-profiles" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.184804 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6739a0a5-d8da-49e3-8131-d2e9607a0455" containerName="collect-profiles" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.185783 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.188086 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-99zgd" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.196171 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww"] Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.197266 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.199865 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.204528 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j"] Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.215559 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-ghqhj"] Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.218561 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.240097 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww"] Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.318780 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7x86\" (UniqueName: \"kubernetes.io/projected/6ec13f6e-cab1-4429-a048-f473cbddcc8f-kube-api-access-g7x86\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.318871 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6ec13f6e-cab1-4429-a048-f473cbddcc8f-dbus-socket\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.318903 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6ec13f6e-cab1-4429-a048-f473cbddcc8f-ovs-socket\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.318964 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6ec13f6e-cab1-4429-a048-f473cbddcc8f-nmstate-lock\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.318996 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5l57\" (UniqueName: \"kubernetes.io/projected/e2cb1036-dcd6-467d-a173-a48732ca9da4-kube-api-access-h5l57\") pod \"nmstate-metrics-fdff9cb8d-8fc4j\" (UID: \"e2cb1036-dcd6-467d-a173-a48732ca9da4\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.319055 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d0610d63-ebfc-43a2-9b50-3ca6cee73d14-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4jxww\" (UID: \"d0610d63-ebfc-43a2-9b50-3ca6cee73d14\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.319079 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjkp2\" (UniqueName: \"kubernetes.io/projected/d0610d63-ebfc-43a2-9b50-3ca6cee73d14-kube-api-access-zjkp2\") pod \"nmstate-webhook-6cdbc54649-4jxww\" (UID: \"d0610d63-ebfc-43a2-9b50-3ca6cee73d14\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.326951 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh"] Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.327832 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.329974 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-zjctc" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.330152 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.330541 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.343744 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh"] Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420226 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ddf1a09e-3019-4a04-b64d-288c8b8cc7d8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-tvcgh\" (UID: \"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420295 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6ec13f6e-cab1-4429-a048-f473cbddcc8f-dbus-socket\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420330 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6ec13f6e-cab1-4429-a048-f473cbddcc8f-ovs-socket\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420368 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dnp5\" (UniqueName: \"kubernetes.io/projected/ddf1a09e-3019-4a04-b64d-288c8b8cc7d8-kube-api-access-2dnp5\") pod \"nmstate-console-plugin-6b874cbd85-tvcgh\" (UID: \"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420405 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddf1a09e-3019-4a04-b64d-288c8b8cc7d8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-tvcgh\" (UID: \"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420438 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6ec13f6e-cab1-4429-a048-f473cbddcc8f-nmstate-lock\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420461 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5l57\" (UniqueName: \"kubernetes.io/projected/e2cb1036-dcd6-467d-a173-a48732ca9da4-kube-api-access-h5l57\") pod \"nmstate-metrics-fdff9cb8d-8fc4j\" (UID: \"e2cb1036-dcd6-467d-a173-a48732ca9da4\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420495 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d0610d63-ebfc-43a2-9b50-3ca6cee73d14-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4jxww\" (UID: \"d0610d63-ebfc-43a2-9b50-3ca6cee73d14\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420520 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjkp2\" (UniqueName: \"kubernetes.io/projected/d0610d63-ebfc-43a2-9b50-3ca6cee73d14-kube-api-access-zjkp2\") pod \"nmstate-webhook-6cdbc54649-4jxww\" (UID: \"d0610d63-ebfc-43a2-9b50-3ca6cee73d14\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.420563 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7x86\" (UniqueName: \"kubernetes.io/projected/6ec13f6e-cab1-4429-a048-f473cbddcc8f-kube-api-access-g7x86\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.421340 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6ec13f6e-cab1-4429-a048-f473cbddcc8f-ovs-socket\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.421417 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6ec13f6e-cab1-4429-a048-f473cbddcc8f-dbus-socket\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.421472 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6ec13f6e-cab1-4429-a048-f473cbddcc8f-nmstate-lock\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.434767 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d0610d63-ebfc-43a2-9b50-3ca6cee73d14-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4jxww\" (UID: \"d0610d63-ebfc-43a2-9b50-3ca6cee73d14\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.439736 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7x86\" (UniqueName: \"kubernetes.io/projected/6ec13f6e-cab1-4429-a048-f473cbddcc8f-kube-api-access-g7x86\") pod \"nmstate-handler-ghqhj\" (UID: \"6ec13f6e-cab1-4429-a048-f473cbddcc8f\") " pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.440124 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjkp2\" (UniqueName: \"kubernetes.io/projected/d0610d63-ebfc-43a2-9b50-3ca6cee73d14-kube-api-access-zjkp2\") pod \"nmstate-webhook-6cdbc54649-4jxww\" (UID: \"d0610d63-ebfc-43a2-9b50-3ca6cee73d14\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.444814 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5l57\" (UniqueName: \"kubernetes.io/projected/e2cb1036-dcd6-467d-a173-a48732ca9da4-kube-api-access-h5l57\") pod \"nmstate-metrics-fdff9cb8d-8fc4j\" (UID: \"e2cb1036-dcd6-467d-a173-a48732ca9da4\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.521664 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ddf1a09e-3019-4a04-b64d-288c8b8cc7d8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-tvcgh\" (UID: \"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.521736 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dnp5\" (UniqueName: \"kubernetes.io/projected/ddf1a09e-3019-4a04-b64d-288c8b8cc7d8-kube-api-access-2dnp5\") pod \"nmstate-console-plugin-6b874cbd85-tvcgh\" (UID: \"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.521769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddf1a09e-3019-4a04-b64d-288c8b8cc7d8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-tvcgh\" (UID: \"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.522849 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ddf1a09e-3019-4a04-b64d-288c8b8cc7d8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-tvcgh\" (UID: \"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.525295 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.528441 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5f7d57f756-tfn2d"] Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.529367 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.532859 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddf1a09e-3019-4a04-b64d-288c8b8cc7d8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-tvcgh\" (UID: \"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.543490 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.545979 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dnp5\" (UniqueName: \"kubernetes.io/projected/ddf1a09e-3019-4a04-b64d-288c8b8cc7d8-kube-api-access-2dnp5\") pod \"nmstate-console-plugin-6b874cbd85-tvcgh\" (UID: \"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.555344 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.598467 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f7d57f756-tfn2d"] Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.622416 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-oauth-serving-cert\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.622666 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-console-serving-cert\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.622685 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-trusted-ca-bundle\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.622715 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-console-oauth-config\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.622737 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnrrx\" (UniqueName: \"kubernetes.io/projected/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-kube-api-access-xnrrx\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.622755 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-console-config\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.622779 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-service-ca\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.645391 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.724562 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-console-serving-cert\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.724602 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-trusted-ca-bundle\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.724638 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-console-oauth-config\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.724665 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnrrx\" (UniqueName: \"kubernetes.io/projected/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-kube-api-access-xnrrx\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.724686 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-console-config\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.725404 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-service-ca\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.725444 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-oauth-serving-cert\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.726162 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-oauth-serving-cert\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.727062 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-console-config\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.727579 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-service-ca\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.730196 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-trusted-ca-bundle\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.730433 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-console-serving-cert\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.732822 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-console-oauth-config\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.742262 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnrrx\" (UniqueName: \"kubernetes.io/projected/9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c-kube-api-access-xnrrx\") pod \"console-5f7d57f756-tfn2d\" (UID: \"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c\") " pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.877391 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh"] Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.922628 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:11 crc kubenswrapper[4770]: I1004 03:15:11.999394 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww"] Oct 04 03:15:12 crc kubenswrapper[4770]: W1004 03:15:12.007416 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0610d63_ebfc_43a2_9b50_3ca6cee73d14.slice/crio-c298dc2cb36a5b0ad01b990167818226c91af45f5fdec12bdea77021fd58ae5e WatchSource:0}: Error finding container c298dc2cb36a5b0ad01b990167818226c91af45f5fdec12bdea77021fd58ae5e: Status 404 returned error can't find the container with id c298dc2cb36a5b0ad01b990167818226c91af45f5fdec12bdea77021fd58ae5e Oct 04 03:15:12 crc kubenswrapper[4770]: I1004 03:15:12.007745 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j"] Oct 04 03:15:12 crc kubenswrapper[4770]: W1004 03:15:12.012406 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2cb1036_dcd6_467d_a173_a48732ca9da4.slice/crio-7e12998351a6253ab4fa1c752b8b69d24cef0135b73d470f66ba4e25e10be0d2 WatchSource:0}: Error finding container 7e12998351a6253ab4fa1c752b8b69d24cef0135b73d470f66ba4e25e10be0d2: Status 404 returned error can't find the container with id 7e12998351a6253ab4fa1c752b8b69d24cef0135b73d470f66ba4e25e10be0d2 Oct 04 03:15:12 crc kubenswrapper[4770]: I1004 03:15:12.099170 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j" event={"ID":"e2cb1036-dcd6-467d-a173-a48732ca9da4","Type":"ContainerStarted","Data":"7e12998351a6253ab4fa1c752b8b69d24cef0135b73d470f66ba4e25e10be0d2"} Oct 04 03:15:12 crc kubenswrapper[4770]: I1004 03:15:12.100420 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" event={"ID":"d0610d63-ebfc-43a2-9b50-3ca6cee73d14","Type":"ContainerStarted","Data":"c298dc2cb36a5b0ad01b990167818226c91af45f5fdec12bdea77021fd58ae5e"} Oct 04 03:15:12 crc kubenswrapper[4770]: I1004 03:15:12.101340 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" event={"ID":"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8","Type":"ContainerStarted","Data":"ad38dd7937cedb3ec79ac75c74e3ba9a608f2e710b475ac108c6e341df4a81e3"} Oct 04 03:15:12 crc kubenswrapper[4770]: I1004 03:15:12.102252 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ghqhj" event={"ID":"6ec13f6e-cab1-4429-a048-f473cbddcc8f","Type":"ContainerStarted","Data":"13df176666ef3724af4714e202208534bbbdd0229238dea7a44445ad656cefed"} Oct 04 03:15:12 crc kubenswrapper[4770]: I1004 03:15:12.142715 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f7d57f756-tfn2d"] Oct 04 03:15:12 crc kubenswrapper[4770]: W1004 03:15:12.148394 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a8cc66f_63c4_49ef_bc9b_1e57e7a1c14c.slice/crio-53ea5ac0af961a16cca95305cb98def1c0d4305a12a14ed32511343bf36431cd WatchSource:0}: Error finding container 53ea5ac0af961a16cca95305cb98def1c0d4305a12a14ed32511343bf36431cd: Status 404 returned error can't find the container with id 53ea5ac0af961a16cca95305cb98def1c0d4305a12a14ed32511343bf36431cd Oct 04 03:15:13 crc kubenswrapper[4770]: I1004 03:15:13.115926 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f7d57f756-tfn2d" event={"ID":"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c","Type":"ContainerStarted","Data":"1c8cb700677ecc12e59ce8c4289e15aed13e2cface090db047e068c08087b329"} Oct 04 03:15:13 crc kubenswrapper[4770]: I1004 03:15:13.116456 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f7d57f756-tfn2d" event={"ID":"9a8cc66f-63c4-49ef-bc9b-1e57e7a1c14c","Type":"ContainerStarted","Data":"53ea5ac0af961a16cca95305cb98def1c0d4305a12a14ed32511343bf36431cd"} Oct 04 03:15:13 crc kubenswrapper[4770]: I1004 03:15:13.157286 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5f7d57f756-tfn2d" podStartSLOduration=2.157256122 podStartE2EDuration="2.157256122s" podCreationTimestamp="2025-10-04 03:15:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:15:13.145649828 +0000 UTC m=+724.437659600" watchObservedRunningTime="2025-10-04 03:15:13.157256122 +0000 UTC m=+724.449265874" Oct 04 03:15:16 crc kubenswrapper[4770]: I1004 03:15:16.137693 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" event={"ID":"ddf1a09e-3019-4a04-b64d-288c8b8cc7d8","Type":"ContainerStarted","Data":"678288676b3f639accb9134210cc9dbbd7772ecde8686368db7f58f7d92b11f1"} Oct 04 03:15:16 crc kubenswrapper[4770]: I1004 03:15:16.139903 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ghqhj" event={"ID":"6ec13f6e-cab1-4429-a048-f473cbddcc8f","Type":"ContainerStarted","Data":"b51c9e032694c9698304f3558008a97a589e95b2b0cbd4bf9d468c405b684a6d"} Oct 04 03:15:16 crc kubenswrapper[4770]: I1004 03:15:16.142266 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" event={"ID":"d0610d63-ebfc-43a2-9b50-3ca6cee73d14","Type":"ContainerStarted","Data":"b2f24143aa58a13fafe5ef16571cc25cae14dbedfe11e61b7f9508db64eb6aaf"} Oct 04 03:15:16 crc kubenswrapper[4770]: I1004 03:15:16.142849 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:16 crc kubenswrapper[4770]: I1004 03:15:16.144563 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j" event={"ID":"e2cb1036-dcd6-467d-a173-a48732ca9da4","Type":"ContainerStarted","Data":"337ee3882e7dce4930f0df95de5cfaf8e26dbdb4137271f7ba61c4a2570ada29"} Oct 04 03:15:16 crc kubenswrapper[4770]: I1004 03:15:16.163200 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-tvcgh" podStartSLOduration=2.015732727 podStartE2EDuration="5.163179275s" podCreationTimestamp="2025-10-04 03:15:11 +0000 UTC" firstStartedPulling="2025-10-04 03:15:11.886826651 +0000 UTC m=+723.178836353" lastFinishedPulling="2025-10-04 03:15:15.034273149 +0000 UTC m=+726.326282901" observedRunningTime="2025-10-04 03:15:16.16108633 +0000 UTC m=+727.453096082" watchObservedRunningTime="2025-10-04 03:15:16.163179275 +0000 UTC m=+727.455188987" Oct 04 03:15:16 crc kubenswrapper[4770]: I1004 03:15:16.212311 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-ghqhj" podStartSLOduration=1.780748682 podStartE2EDuration="5.212247046s" podCreationTimestamp="2025-10-04 03:15:11 +0000 UTC" firstStartedPulling="2025-10-04 03:15:11.614120321 +0000 UTC m=+722.906130033" lastFinishedPulling="2025-10-04 03:15:15.045618645 +0000 UTC m=+726.337628397" observedRunningTime="2025-10-04 03:15:16.188918137 +0000 UTC m=+727.480927929" watchObservedRunningTime="2025-10-04 03:15:16.212247046 +0000 UTC m=+727.504256758" Oct 04 03:15:16 crc kubenswrapper[4770]: I1004 03:15:16.213894 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" podStartSLOduration=2.179764239 podStartE2EDuration="5.213861818s" podCreationTimestamp="2025-10-04 03:15:11 +0000 UTC" firstStartedPulling="2025-10-04 03:15:12.010737736 +0000 UTC m=+723.302747448" lastFinishedPulling="2025-10-04 03:15:15.044835305 +0000 UTC m=+726.336845027" observedRunningTime="2025-10-04 03:15:16.211428915 +0000 UTC m=+727.503438627" watchObservedRunningTime="2025-10-04 03:15:16.213861818 +0000 UTC m=+727.505871550" Oct 04 03:15:16 crc kubenswrapper[4770]: I1004 03:15:16.557126 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:18 crc kubenswrapper[4770]: I1004 03:15:18.167188 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j" event={"ID":"e2cb1036-dcd6-467d-a173-a48732ca9da4","Type":"ContainerStarted","Data":"34b653207d62ec626429079ae6dbc335788f43c498f9801d02db0a087dded736"} Oct 04 03:15:21 crc kubenswrapper[4770]: I1004 03:15:21.595864 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-ghqhj" Oct 04 03:15:21 crc kubenswrapper[4770]: I1004 03:15:21.622813 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8fc4j" podStartSLOduration=4.741386182 podStartE2EDuration="10.622778533s" podCreationTimestamp="2025-10-04 03:15:11 +0000 UTC" firstStartedPulling="2025-10-04 03:15:12.015557662 +0000 UTC m=+723.307567374" lastFinishedPulling="2025-10-04 03:15:17.896949973 +0000 UTC m=+729.188959725" observedRunningTime="2025-10-04 03:15:18.201165556 +0000 UTC m=+729.493175278" watchObservedRunningTime="2025-10-04 03:15:21.622778533 +0000 UTC m=+732.914788275" Oct 04 03:15:21 crc kubenswrapper[4770]: I1004 03:15:21.923126 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:21 crc kubenswrapper[4770]: I1004 03:15:21.923205 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:21 crc kubenswrapper[4770]: I1004 03:15:21.928886 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:22 crc kubenswrapper[4770]: I1004 03:15:22.211699 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5f7d57f756-tfn2d" Oct 04 03:15:22 crc kubenswrapper[4770]: I1004 03:15:22.289734 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xwhz4"] Oct 04 03:15:31 crc kubenswrapper[4770]: I1004 03:15:31.554674 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4jxww" Oct 04 03:15:34 crc kubenswrapper[4770]: I1004 03:15:34.609606 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xkkrp"] Oct 04 03:15:34 crc kubenswrapper[4770]: I1004 03:15:34.611563 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" podUID="a1483904-9014-4f9d-944b-3612fbd3b3d9" containerName="controller-manager" containerID="cri-o://f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58" gracePeriod=30 Oct 04 03:15:34 crc kubenswrapper[4770]: I1004 03:15:34.716997 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9"] Oct 04 03:15:34 crc kubenswrapper[4770]: I1004 03:15:34.717286 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" podUID="c4c445d6-f62c-4d63-8d18-35441eebeb49" containerName="route-controller-manager" containerID="cri-o://f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28" gracePeriod=30 Oct 04 03:15:34 crc kubenswrapper[4770]: I1004 03:15:34.984720 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.005774 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frncq\" (UniqueName: \"kubernetes.io/projected/a1483904-9014-4f9d-944b-3612fbd3b3d9-kube-api-access-frncq\") pod \"a1483904-9014-4f9d-944b-3612fbd3b3d9\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.005917 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-client-ca\") pod \"a1483904-9014-4f9d-944b-3612fbd3b3d9\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.006036 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-config\") pod \"a1483904-9014-4f9d-944b-3612fbd3b3d9\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.006071 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1483904-9014-4f9d-944b-3612fbd3b3d9-serving-cert\") pod \"a1483904-9014-4f9d-944b-3612fbd3b3d9\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.006099 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-proxy-ca-bundles\") pod \"a1483904-9014-4f9d-944b-3612fbd3b3d9\" (UID: \"a1483904-9014-4f9d-944b-3612fbd3b3d9\") " Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.007209 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a1483904-9014-4f9d-944b-3612fbd3b3d9" (UID: "a1483904-9014-4f9d-944b-3612fbd3b3d9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.007223 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-client-ca" (OuterVolumeSpecName: "client-ca") pod "a1483904-9014-4f9d-944b-3612fbd3b3d9" (UID: "a1483904-9014-4f9d-944b-3612fbd3b3d9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.007341 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-config" (OuterVolumeSpecName: "config") pod "a1483904-9014-4f9d-944b-3612fbd3b3d9" (UID: "a1483904-9014-4f9d-944b-3612fbd3b3d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.015561 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1483904-9014-4f9d-944b-3612fbd3b3d9-kube-api-access-frncq" (OuterVolumeSpecName: "kube-api-access-frncq") pod "a1483904-9014-4f9d-944b-3612fbd3b3d9" (UID: "a1483904-9014-4f9d-944b-3612fbd3b3d9"). InnerVolumeSpecName "kube-api-access-frncq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.021619 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1483904-9014-4f9d-944b-3612fbd3b3d9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a1483904-9014-4f9d-944b-3612fbd3b3d9" (UID: "a1483904-9014-4f9d-944b-3612fbd3b3d9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.045131 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.107584 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c445d6-f62c-4d63-8d18-35441eebeb49-serving-cert\") pod \"c4c445d6-f62c-4d63-8d18-35441eebeb49\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.107651 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvfs7\" (UniqueName: \"kubernetes.io/projected/c4c445d6-f62c-4d63-8d18-35441eebeb49-kube-api-access-vvfs7\") pod \"c4c445d6-f62c-4d63-8d18-35441eebeb49\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.107724 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-config\") pod \"c4c445d6-f62c-4d63-8d18-35441eebeb49\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.107761 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-client-ca\") pod \"c4c445d6-f62c-4d63-8d18-35441eebeb49\" (UID: \"c4c445d6-f62c-4d63-8d18-35441eebeb49\") " Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.107984 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.108017 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1483904-9014-4f9d-944b-3612fbd3b3d9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.108027 4770 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.108038 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frncq\" (UniqueName: \"kubernetes.io/projected/a1483904-9014-4f9d-944b-3612fbd3b3d9-kube-api-access-frncq\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.108046 4770 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a1483904-9014-4f9d-944b-3612fbd3b3d9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.108934 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-client-ca" (OuterVolumeSpecName: "client-ca") pod "c4c445d6-f62c-4d63-8d18-35441eebeb49" (UID: "c4c445d6-f62c-4d63-8d18-35441eebeb49"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.109123 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-config" (OuterVolumeSpecName: "config") pod "c4c445d6-f62c-4d63-8d18-35441eebeb49" (UID: "c4c445d6-f62c-4d63-8d18-35441eebeb49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.111223 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c445d6-f62c-4d63-8d18-35441eebeb49-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c4c445d6-f62c-4d63-8d18-35441eebeb49" (UID: "c4c445d6-f62c-4d63-8d18-35441eebeb49"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.111445 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4c445d6-f62c-4d63-8d18-35441eebeb49-kube-api-access-vvfs7" (OuterVolumeSpecName: "kube-api-access-vvfs7") pod "c4c445d6-f62c-4d63-8d18-35441eebeb49" (UID: "c4c445d6-f62c-4d63-8d18-35441eebeb49"). InnerVolumeSpecName "kube-api-access-vvfs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.209166 4770 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c445d6-f62c-4d63-8d18-35441eebeb49-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.209212 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvfs7\" (UniqueName: \"kubernetes.io/projected/c4c445d6-f62c-4d63-8d18-35441eebeb49-kube-api-access-vvfs7\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.209226 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.209234 4770 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4c445d6-f62c-4d63-8d18-35441eebeb49-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.306292 4770 generic.go:334] "Generic (PLEG): container finished" podID="c4c445d6-f62c-4d63-8d18-35441eebeb49" containerID="f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28" exitCode=0 Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.306373 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.306389 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" event={"ID":"c4c445d6-f62c-4d63-8d18-35441eebeb49","Type":"ContainerDied","Data":"f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28"} Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.306467 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9" event={"ID":"c4c445d6-f62c-4d63-8d18-35441eebeb49","Type":"ContainerDied","Data":"b2ffb19386f5c458934aa45f349121fd237e7cdfe843c2198c7e1bd5c8c576b0"} Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.306491 4770 scope.go:117] "RemoveContainer" containerID="f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.308992 4770 generic.go:334] "Generic (PLEG): container finished" podID="a1483904-9014-4f9d-944b-3612fbd3b3d9" containerID="f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58" exitCode=0 Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.309108 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" event={"ID":"a1483904-9014-4f9d-944b-3612fbd3b3d9","Type":"ContainerDied","Data":"f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58"} Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.309172 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" event={"ID":"a1483904-9014-4f9d-944b-3612fbd3b3d9","Type":"ContainerDied","Data":"5398ab6bbf37f7249c595d6be58131a5b83f22fbc7184e2b560e2e47db153171"} Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.309365 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xkkrp" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.349059 4770 scope.go:117] "RemoveContainer" containerID="f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28" Oct 04 03:15:35 crc kubenswrapper[4770]: E1004 03:15:35.350865 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28\": container with ID starting with f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28 not found: ID does not exist" containerID="f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.350919 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28"} err="failed to get container status \"f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28\": rpc error: code = NotFound desc = could not find container \"f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28\": container with ID starting with f0000098c634496f4a959a4803452e4f6f6f3a353864007dfab8f521955c2f28 not found: ID does not exist" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.350955 4770 scope.go:117] "RemoveContainer" containerID="f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.357934 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9"] Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.361487 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-jq7w9"] Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.372917 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xkkrp"] Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.375427 4770 scope.go:117] "RemoveContainer" containerID="f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58" Oct 04 03:15:35 crc kubenswrapper[4770]: E1004 03:15:35.375891 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58\": container with ID starting with f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58 not found: ID does not exist" containerID="f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.375924 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58"} err="failed to get container status \"f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58\": rpc error: code = NotFound desc = could not find container \"f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58\": container with ID starting with f59ac0f72035015d74626065b8079aedb8edea98cd581dad93cf3cda05c76e58 not found: ID does not exist" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.379047 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xkkrp"] Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.681676 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1483904-9014-4f9d-944b-3612fbd3b3d9" path="/var/lib/kubelet/pods/a1483904-9014-4f9d-944b-3612fbd3b3d9/volumes" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.682744 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4c445d6-f62c-4d63-8d18-35441eebeb49" path="/var/lib/kubelet/pods/c4c445d6-f62c-4d63-8d18-35441eebeb49/volumes" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.997897 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl"] Oct 04 03:15:35 crc kubenswrapper[4770]: E1004 03:15:35.998426 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1483904-9014-4f9d-944b-3612fbd3b3d9" containerName="controller-manager" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.998459 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1483904-9014-4f9d-944b-3612fbd3b3d9" containerName="controller-manager" Oct 04 03:15:35 crc kubenswrapper[4770]: E1004 03:15:35.998490 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c445d6-f62c-4d63-8d18-35441eebeb49" containerName="route-controller-manager" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.998507 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c445d6-f62c-4d63-8d18-35441eebeb49" containerName="route-controller-manager" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.998698 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4c445d6-f62c-4d63-8d18-35441eebeb49" containerName="route-controller-manager" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.998743 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1483904-9014-4f9d-944b-3612fbd3b3d9" containerName="controller-manager" Oct 04 03:15:35 crc kubenswrapper[4770]: I1004 03:15:35.999608 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.001377 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-77796db85d-ww2vj"] Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.002346 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.003763 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.003936 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.004061 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.004227 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.003763 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.004648 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.010891 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.011276 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.011622 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.011771 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.011958 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.012089 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl"] Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.012127 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.023573 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.031084 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77796db85d-ww2vj"] Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.122538 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhhbk\" (UniqueName: \"kubernetes.io/projected/e06452e3-8e6c-4858-8f99-d321b8a326c9-kube-api-access-nhhbk\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.123211 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06452e3-8e6c-4858-8f99-d321b8a326c9-config\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.123641 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18db710f-f644-4874-8516-5a2dd96b4107-serving-cert\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.123791 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e06452e3-8e6c-4858-8f99-d321b8a326c9-serving-cert\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.123915 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e06452e3-8e6c-4858-8f99-d321b8a326c9-proxy-ca-bundles\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.124069 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18db710f-f644-4874-8516-5a2dd96b4107-config\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.124213 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e06452e3-8e6c-4858-8f99-d321b8a326c9-client-ca\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.124366 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18db710f-f644-4874-8516-5a2dd96b4107-client-ca\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.124481 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glw55\" (UniqueName: \"kubernetes.io/projected/18db710f-f644-4874-8516-5a2dd96b4107-kube-api-access-glw55\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.225988 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e06452e3-8e6c-4858-8f99-d321b8a326c9-client-ca\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.226562 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18db710f-f644-4874-8516-5a2dd96b4107-client-ca\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.226767 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glw55\" (UniqueName: \"kubernetes.io/projected/18db710f-f644-4874-8516-5a2dd96b4107-kube-api-access-glw55\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.226931 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhhbk\" (UniqueName: \"kubernetes.io/projected/e06452e3-8e6c-4858-8f99-d321b8a326c9-kube-api-access-nhhbk\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.227066 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e06452e3-8e6c-4858-8f99-d321b8a326c9-client-ca\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.227707 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06452e3-8e6c-4858-8f99-d321b8a326c9-config\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.228097 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18db710f-f644-4874-8516-5a2dd96b4107-serving-cert\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.227961 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18db710f-f644-4874-8516-5a2dd96b4107-client-ca\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.228272 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e06452e3-8e6c-4858-8f99-d321b8a326c9-serving-cert\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.228513 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e06452e3-8e6c-4858-8f99-d321b8a326c9-proxy-ca-bundles\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.228865 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18db710f-f644-4874-8516-5a2dd96b4107-config\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.229554 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e06452e3-8e6c-4858-8f99-d321b8a326c9-config\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.229765 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e06452e3-8e6c-4858-8f99-d321b8a326c9-proxy-ca-bundles\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.230668 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18db710f-f644-4874-8516-5a2dd96b4107-config\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.237646 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18db710f-f644-4874-8516-5a2dd96b4107-serving-cert\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.238267 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e06452e3-8e6c-4858-8f99-d321b8a326c9-serving-cert\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.256063 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhhbk\" (UniqueName: \"kubernetes.io/projected/e06452e3-8e6c-4858-8f99-d321b8a326c9-kube-api-access-nhhbk\") pod \"controller-manager-77796db85d-ww2vj\" (UID: \"e06452e3-8e6c-4858-8f99-d321b8a326c9\") " pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.257738 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glw55\" (UniqueName: \"kubernetes.io/projected/18db710f-f644-4874-8516-5a2dd96b4107-kube-api-access-glw55\") pod \"route-controller-manager-644fbc58-bqxpl\" (UID: \"18db710f-f644-4874-8516-5a2dd96b4107\") " pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.333514 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.343125 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.563029 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77796db85d-ww2vj"] Oct 04 03:15:36 crc kubenswrapper[4770]: I1004 03:15:36.633105 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl"] Oct 04 03:15:36 crc kubenswrapper[4770]: W1004 03:15:36.638732 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18db710f_f644_4874_8516_5a2dd96b4107.slice/crio-41a94a349bab64251e6f4855b685b6c428e464404baef85a551e2d60a1641671 WatchSource:0}: Error finding container 41a94a349bab64251e6f4855b685b6c428e464404baef85a551e2d60a1641671: Status 404 returned error can't find the container with id 41a94a349bab64251e6f4855b685b6c428e464404baef85a551e2d60a1641671 Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.323202 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" event={"ID":"e06452e3-8e6c-4858-8f99-d321b8a326c9","Type":"ContainerStarted","Data":"766715989fb1644a231188e015d73a836e7a624c36f511ec9ed513dc6d145198"} Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.323552 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" event={"ID":"e06452e3-8e6c-4858-8f99-d321b8a326c9","Type":"ContainerStarted","Data":"1b5b3526b2e0aca65479dcec6b6ab92773fa38632550ae82dbebac67a9b62438"} Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.323574 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.324653 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" event={"ID":"18db710f-f644-4874-8516-5a2dd96b4107","Type":"ContainerStarted","Data":"f930ca72ae0be5d0dfb1aef264802cdcaf0ce5662709f2992e245d246148b8ec"} Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.324717 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" event={"ID":"18db710f-f644-4874-8516-5a2dd96b4107","Type":"ContainerStarted","Data":"41a94a349bab64251e6f4855b685b6c428e464404baef85a551e2d60a1641671"} Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.324879 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.328071 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.329854 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.340169 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-77796db85d-ww2vj" podStartSLOduration=3.340141536 podStartE2EDuration="3.340141536s" podCreationTimestamp="2025-10-04 03:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:15:37.339802037 +0000 UTC m=+748.631811759" watchObservedRunningTime="2025-10-04 03:15:37.340141536 +0000 UTC m=+748.632151248" Oct 04 03:15:37 crc kubenswrapper[4770]: I1004 03:15:37.361777 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-644fbc58-bqxpl" podStartSLOduration=3.36174989 podStartE2EDuration="3.36174989s" podCreationTimestamp="2025-10-04 03:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:15:37.358170597 +0000 UTC m=+748.650180309" watchObservedRunningTime="2025-10-04 03:15:37.36174989 +0000 UTC m=+748.653759602" Oct 04 03:15:43 crc kubenswrapper[4770]: I1004 03:15:43.359203 4770 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.346262 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-xwhz4" podUID="2b45833c-2e11-428e-8131-1ea4676a794f" containerName="console" containerID="cri-o://5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556" gracePeriod=15 Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.516693 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7"] Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.518421 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.528950 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.529780 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7"] Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.609059 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pblb9\" (UniqueName: \"kubernetes.io/projected/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-kube-api-access-pblb9\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.609470 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.609541 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.711786 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.712015 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pblb9\" (UniqueName: \"kubernetes.io/projected/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-kube-api-access-pblb9\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.712055 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.712347 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.712353 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.730786 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pblb9\" (UniqueName: \"kubernetes.io/projected/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-kube-api-access-pblb9\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.772394 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xwhz4_2b45833c-2e11-428e-8131-1ea4676a794f/console/0.log" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.772480 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.836134 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.914735 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-oauth-config\") pod \"2b45833c-2e11-428e-8131-1ea4676a794f\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.914833 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-serving-cert\") pod \"2b45833c-2e11-428e-8131-1ea4676a794f\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.914875 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-trusted-ca-bundle\") pod \"2b45833c-2e11-428e-8131-1ea4676a794f\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.914904 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-console-config\") pod \"2b45833c-2e11-428e-8131-1ea4676a794f\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.914944 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjhlh\" (UniqueName: \"kubernetes.io/projected/2b45833c-2e11-428e-8131-1ea4676a794f-kube-api-access-vjhlh\") pod \"2b45833c-2e11-428e-8131-1ea4676a794f\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.914978 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-oauth-serving-cert\") pod \"2b45833c-2e11-428e-8131-1ea4676a794f\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.915037 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-service-ca\") pod \"2b45833c-2e11-428e-8131-1ea4676a794f\" (UID: \"2b45833c-2e11-428e-8131-1ea4676a794f\") " Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.917637 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-service-ca" (OuterVolumeSpecName: "service-ca") pod "2b45833c-2e11-428e-8131-1ea4676a794f" (UID: "2b45833c-2e11-428e-8131-1ea4676a794f"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.917855 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-console-config" (OuterVolumeSpecName: "console-config") pod "2b45833c-2e11-428e-8131-1ea4676a794f" (UID: "2b45833c-2e11-428e-8131-1ea4676a794f"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.918555 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "2b45833c-2e11-428e-8131-1ea4676a794f" (UID: "2b45833c-2e11-428e-8131-1ea4676a794f"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.918666 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "2b45833c-2e11-428e-8131-1ea4676a794f" (UID: "2b45833c-2e11-428e-8131-1ea4676a794f"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.922162 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "2b45833c-2e11-428e-8131-1ea4676a794f" (UID: "2b45833c-2e11-428e-8131-1ea4676a794f"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.924130 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b45833c-2e11-428e-8131-1ea4676a794f-kube-api-access-vjhlh" (OuterVolumeSpecName: "kube-api-access-vjhlh") pod "2b45833c-2e11-428e-8131-1ea4676a794f" (UID: "2b45833c-2e11-428e-8131-1ea4676a794f"). InnerVolumeSpecName "kube-api-access-vjhlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:47 crc kubenswrapper[4770]: I1004 03:15:47.924664 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "2b45833c-2e11-428e-8131-1ea4676a794f" (UID: "2b45833c-2e11-428e-8131-1ea4676a794f"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.016545 4770 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.016629 4770 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.016654 4770 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b45833c-2e11-428e-8131-1ea4676a794f-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.016676 4770 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.016699 4770 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.016716 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjhlh\" (UniqueName: \"kubernetes.io/projected/2b45833c-2e11-428e-8131-1ea4676a794f-kube-api-access-vjhlh\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.016737 4770 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b45833c-2e11-428e-8131-1ea4676a794f-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.253115 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7"] Oct 04 03:15:48 crc kubenswrapper[4770]: W1004 03:15:48.256725 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d6f0476_54a4_43f3_8fe8_737fd8c4dc17.slice/crio-7c02348d095589bed7200ece0acfbc537571451fd3285f093717edde2972b6ca WatchSource:0}: Error finding container 7c02348d095589bed7200ece0acfbc537571451fd3285f093717edde2972b6ca: Status 404 returned error can't find the container with id 7c02348d095589bed7200ece0acfbc537571451fd3285f093717edde2972b6ca Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.404871 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xwhz4_2b45833c-2e11-428e-8131-1ea4676a794f/console/0.log" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.404936 4770 generic.go:334] "Generic (PLEG): container finished" podID="2b45833c-2e11-428e-8131-1ea4676a794f" containerID="5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556" exitCode=2 Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.405064 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xwhz4" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.405072 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xwhz4" event={"ID":"2b45833c-2e11-428e-8131-1ea4676a794f","Type":"ContainerDied","Data":"5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556"} Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.405205 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xwhz4" event={"ID":"2b45833c-2e11-428e-8131-1ea4676a794f","Type":"ContainerDied","Data":"9f3292fbfd3849da26529499e7e7fe3384b2e23d9bb336a97a84721af063c3f7"} Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.405225 4770 scope.go:117] "RemoveContainer" containerID="5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.407136 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" event={"ID":"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17","Type":"ContainerStarted","Data":"7c02348d095589bed7200ece0acfbc537571451fd3285f093717edde2972b6ca"} Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.429085 4770 scope.go:117] "RemoveContainer" containerID="5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556" Oct 04 03:15:48 crc kubenswrapper[4770]: E1004 03:15:48.429955 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556\": container with ID starting with 5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556 not found: ID does not exist" containerID="5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.430105 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556"} err="failed to get container status \"5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556\": rpc error: code = NotFound desc = could not find container \"5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556\": container with ID starting with 5ee51ebe4e70f42ffac21b6d2e093444df200bd73a7e06207a5094b5b37cc556 not found: ID does not exist" Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.438413 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xwhz4"] Oct 04 03:15:48 crc kubenswrapper[4770]: I1004 03:15:48.446344 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-xwhz4"] Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.418469 4770 generic.go:334] "Generic (PLEG): container finished" podID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerID="792ee8c50ac41064f0fa42dafd07b3efec2cd26bbc68c7562cc80ee32f8d0864" exitCode=0 Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.418587 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" event={"ID":"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17","Type":"ContainerDied","Data":"792ee8c50ac41064f0fa42dafd07b3efec2cd26bbc68c7562cc80ee32f8d0864"} Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.693402 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b45833c-2e11-428e-8131-1ea4676a794f" path="/var/lib/kubelet/pods/2b45833c-2e11-428e-8131-1ea4676a794f/volumes" Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.826471 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6bwp7"] Oct 04 03:15:49 crc kubenswrapper[4770]: E1004 03:15:49.826916 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b45833c-2e11-428e-8131-1ea4676a794f" containerName="console" Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.826941 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b45833c-2e11-428e-8131-1ea4676a794f" containerName="console" Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.827215 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b45833c-2e11-428e-8131-1ea4676a794f" containerName="console" Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.829049 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.840000 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6bwp7"] Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.942651 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-utilities\") pod \"redhat-operators-6bwp7\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.942708 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbxkq\" (UniqueName: \"kubernetes.io/projected/34745ebf-06a4-4019-beb3-4fd2e2855150-kube-api-access-vbxkq\") pod \"redhat-operators-6bwp7\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:49 crc kubenswrapper[4770]: I1004 03:15:49.942744 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-catalog-content\") pod \"redhat-operators-6bwp7\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:50 crc kubenswrapper[4770]: I1004 03:15:50.043705 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-utilities\") pod \"redhat-operators-6bwp7\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:50 crc kubenswrapper[4770]: I1004 03:15:50.043764 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbxkq\" (UniqueName: \"kubernetes.io/projected/34745ebf-06a4-4019-beb3-4fd2e2855150-kube-api-access-vbxkq\") pod \"redhat-operators-6bwp7\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:50 crc kubenswrapper[4770]: I1004 03:15:50.043799 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-catalog-content\") pod \"redhat-operators-6bwp7\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:50 crc kubenswrapper[4770]: I1004 03:15:50.044430 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-catalog-content\") pod \"redhat-operators-6bwp7\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:50 crc kubenswrapper[4770]: I1004 03:15:50.044419 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-utilities\") pod \"redhat-operators-6bwp7\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:50 crc kubenswrapper[4770]: I1004 03:15:50.080518 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbxkq\" (UniqueName: \"kubernetes.io/projected/34745ebf-06a4-4019-beb3-4fd2e2855150-kube-api-access-vbxkq\") pod \"redhat-operators-6bwp7\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:50 crc kubenswrapper[4770]: I1004 03:15:50.152514 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:15:50 crc kubenswrapper[4770]: I1004 03:15:50.575648 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6bwp7"] Oct 04 03:15:51 crc kubenswrapper[4770]: I1004 03:15:51.444492 4770 generic.go:334] "Generic (PLEG): container finished" podID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerID="9a8787588e1994c54569ffc77a5059b91b88e925d5a8aef307a0638ef0d748f3" exitCode=0 Oct 04 03:15:51 crc kubenswrapper[4770]: I1004 03:15:51.444679 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" event={"ID":"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17","Type":"ContainerDied","Data":"9a8787588e1994c54569ffc77a5059b91b88e925d5a8aef307a0638ef0d748f3"} Oct 04 03:15:51 crc kubenswrapper[4770]: I1004 03:15:51.451553 4770 generic.go:334] "Generic (PLEG): container finished" podID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerID="722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb" exitCode=0 Oct 04 03:15:51 crc kubenswrapper[4770]: I1004 03:15:51.451613 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bwp7" event={"ID":"34745ebf-06a4-4019-beb3-4fd2e2855150","Type":"ContainerDied","Data":"722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb"} Oct 04 03:15:51 crc kubenswrapper[4770]: I1004 03:15:51.451653 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bwp7" event={"ID":"34745ebf-06a4-4019-beb3-4fd2e2855150","Type":"ContainerStarted","Data":"7440888233da33cd738e4032f36180bac327ddda35333d14607ce12425eee1a2"} Oct 04 03:15:52 crc kubenswrapper[4770]: I1004 03:15:52.465421 4770 generic.go:334] "Generic (PLEG): container finished" podID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerID="cc22807b4f2b996f616f7c6198ed89b70f6ffeae8fde0ba46cc945dbe3603e2b" exitCode=0 Oct 04 03:15:52 crc kubenswrapper[4770]: I1004 03:15:52.465498 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" event={"ID":"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17","Type":"ContainerDied","Data":"cc22807b4f2b996f616f7c6198ed89b70f6ffeae8fde0ba46cc945dbe3603e2b"} Oct 04 03:15:53 crc kubenswrapper[4770]: I1004 03:15:53.476148 4770 generic.go:334] "Generic (PLEG): container finished" podID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerID="f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5" exitCode=0 Oct 04 03:15:53 crc kubenswrapper[4770]: I1004 03:15:53.476295 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bwp7" event={"ID":"34745ebf-06a4-4019-beb3-4fd2e2855150","Type":"ContainerDied","Data":"f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5"} Oct 04 03:15:53 crc kubenswrapper[4770]: I1004 03:15:53.875389 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:53 crc kubenswrapper[4770]: I1004 03:15:53.997585 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pblb9\" (UniqueName: \"kubernetes.io/projected/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-kube-api-access-pblb9\") pod \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " Oct 04 03:15:53 crc kubenswrapper[4770]: I1004 03:15:53.997677 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-util\") pod \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " Oct 04 03:15:53 crc kubenswrapper[4770]: I1004 03:15:53.997698 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-bundle\") pod \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\" (UID: \"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17\") " Oct 04 03:15:53 crc kubenswrapper[4770]: I1004 03:15:53.999182 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-bundle" (OuterVolumeSpecName: "bundle") pod "7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" (UID: "7d6f0476-54a4-43f3-8fe8-737fd8c4dc17"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.008422 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-kube-api-access-pblb9" (OuterVolumeSpecName: "kube-api-access-pblb9") pod "7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" (UID: "7d6f0476-54a4-43f3-8fe8-737fd8c4dc17"). InnerVolumeSpecName "kube-api-access-pblb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.100045 4770 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.100096 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pblb9\" (UniqueName: \"kubernetes.io/projected/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-kube-api-access-pblb9\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.323260 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-util" (OuterVolumeSpecName: "util") pod "7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" (UID: "7d6f0476-54a4-43f3-8fe8-737fd8c4dc17"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.405148 4770 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d6f0476-54a4-43f3-8fe8-737fd8c4dc17-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.488790 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" event={"ID":"7d6f0476-54a4-43f3-8fe8-737fd8c4dc17","Type":"ContainerDied","Data":"7c02348d095589bed7200ece0acfbc537571451fd3285f093717edde2972b6ca"} Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.488866 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c02348d095589bed7200ece0acfbc537571451fd3285f093717edde2972b6ca" Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.488994 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7" Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.493979 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bwp7" event={"ID":"34745ebf-06a4-4019-beb3-4fd2e2855150","Type":"ContainerStarted","Data":"db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f"} Oct 04 03:15:54 crc kubenswrapper[4770]: I1004 03:15:54.514086 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6bwp7" podStartSLOduration=3.09044386 podStartE2EDuration="5.51406694s" podCreationTimestamp="2025-10-04 03:15:49 +0000 UTC" firstStartedPulling="2025-10-04 03:15:51.469270442 +0000 UTC m=+762.761280154" lastFinishedPulling="2025-10-04 03:15:53.892893512 +0000 UTC m=+765.184903234" observedRunningTime="2025-10-04 03:15:54.51251911 +0000 UTC m=+765.804528842" watchObservedRunningTime="2025-10-04 03:15:54.51406694 +0000 UTC m=+765.806076652" Oct 04 03:16:00 crc kubenswrapper[4770]: I1004 03:16:00.153197 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:16:00 crc kubenswrapper[4770]: I1004 03:16:00.154329 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:16:00 crc kubenswrapper[4770]: I1004 03:16:00.211099 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:16:00 crc kubenswrapper[4770]: I1004 03:16:00.571991 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:16:01 crc kubenswrapper[4770]: I1004 03:16:01.999253 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6bwp7"] Oct 04 03:16:02 crc kubenswrapper[4770]: I1004 03:16:02.544979 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6bwp7" podUID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerName="registry-server" containerID="cri-o://db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f" gracePeriod=2 Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.063677 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.137178 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-catalog-content\") pod \"34745ebf-06a4-4019-beb3-4fd2e2855150\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.137264 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbxkq\" (UniqueName: \"kubernetes.io/projected/34745ebf-06a4-4019-beb3-4fd2e2855150-kube-api-access-vbxkq\") pod \"34745ebf-06a4-4019-beb3-4fd2e2855150\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.137325 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-utilities\") pod \"34745ebf-06a4-4019-beb3-4fd2e2855150\" (UID: \"34745ebf-06a4-4019-beb3-4fd2e2855150\") " Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.138352 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-utilities" (OuterVolumeSpecName: "utilities") pod "34745ebf-06a4-4019-beb3-4fd2e2855150" (UID: "34745ebf-06a4-4019-beb3-4fd2e2855150"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.148515 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34745ebf-06a4-4019-beb3-4fd2e2855150-kube-api-access-vbxkq" (OuterVolumeSpecName: "kube-api-access-vbxkq") pod "34745ebf-06a4-4019-beb3-4fd2e2855150" (UID: "34745ebf-06a4-4019-beb3-4fd2e2855150"). InnerVolumeSpecName "kube-api-access-vbxkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.238772 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbxkq\" (UniqueName: \"kubernetes.io/projected/34745ebf-06a4-4019-beb3-4fd2e2855150-kube-api-access-vbxkq\") on node \"crc\" DevicePath \"\"" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.238812 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.555383 4770 generic.go:334] "Generic (PLEG): container finished" podID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerID="db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f" exitCode=0 Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.555464 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bwp7" event={"ID":"34745ebf-06a4-4019-beb3-4fd2e2855150","Type":"ContainerDied","Data":"db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f"} Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.555530 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bwp7" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.555549 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bwp7" event={"ID":"34745ebf-06a4-4019-beb3-4fd2e2855150","Type":"ContainerDied","Data":"7440888233da33cd738e4032f36180bac327ddda35333d14607ce12425eee1a2"} Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.555582 4770 scope.go:117] "RemoveContainer" containerID="db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.577541 4770 scope.go:117] "RemoveContainer" containerID="f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.600912 4770 scope.go:117] "RemoveContainer" containerID="722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.625470 4770 scope.go:117] "RemoveContainer" containerID="db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f" Oct 04 03:16:03 crc kubenswrapper[4770]: E1004 03:16:03.626089 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f\": container with ID starting with db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f not found: ID does not exist" containerID="db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.626136 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f"} err="failed to get container status \"db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f\": rpc error: code = NotFound desc = could not find container \"db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f\": container with ID starting with db6432a08833c84203e34ae5531ab50f262bc76bd93d5ba2ca299e85bd7b228f not found: ID does not exist" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.626167 4770 scope.go:117] "RemoveContainer" containerID="f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5" Oct 04 03:16:03 crc kubenswrapper[4770]: E1004 03:16:03.626631 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5\": container with ID starting with f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5 not found: ID does not exist" containerID="f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.626669 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5"} err="failed to get container status \"f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5\": rpc error: code = NotFound desc = could not find container \"f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5\": container with ID starting with f5a36a211fa9856a252cc8b448e2f9bbf845e8788a8395f2fffcf08996005fc5 not found: ID does not exist" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.626696 4770 scope.go:117] "RemoveContainer" containerID="722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb" Oct 04 03:16:03 crc kubenswrapper[4770]: E1004 03:16:03.627258 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb\": container with ID starting with 722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb not found: ID does not exist" containerID="722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb" Oct 04 03:16:03 crc kubenswrapper[4770]: I1004 03:16:03.627288 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb"} err="failed to get container status \"722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb\": rpc error: code = NotFound desc = could not find container \"722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb\": container with ID starting with 722253e92a37c30a93afcdaf054021f164f75b4b885b9c1884ba949451ad57bb not found: ID does not exist" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.142565 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34745ebf-06a4-4019-beb3-4fd2e2855150" (UID: "34745ebf-06a4-4019-beb3-4fd2e2855150"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.152038 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34745ebf-06a4-4019-beb3-4fd2e2855150-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.195819 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6bwp7"] Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.199309 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6bwp7"] Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.532556 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db"] Oct 04 03:16:04 crc kubenswrapper[4770]: E1004 03:16:04.532835 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerName="extract-content" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.532855 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerName="extract-content" Oct 04 03:16:04 crc kubenswrapper[4770]: E1004 03:16:04.532870 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerName="registry-server" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.532880 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerName="registry-server" Oct 04 03:16:04 crc kubenswrapper[4770]: E1004 03:16:04.532889 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerName="pull" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.532896 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerName="pull" Oct 04 03:16:04 crc kubenswrapper[4770]: E1004 03:16:04.532906 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerName="util" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.532912 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerName="util" Oct 04 03:16:04 crc kubenswrapper[4770]: E1004 03:16:04.532925 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerName="extract-utilities" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.532933 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerName="extract-utilities" Oct 04 03:16:04 crc kubenswrapper[4770]: E1004 03:16:04.532946 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerName="extract" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.532954 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerName="extract" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.533131 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="34745ebf-06a4-4019-beb3-4fd2e2855150" containerName="registry-server" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.533157 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d6f0476-54a4-43f3-8fe8-737fd8c4dc17" containerName="extract" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.533684 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.535594 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.535665 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.536515 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.537500 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.548144 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db"] Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.554159 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-49slr" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.673563 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e75166b5-2b64-4717-9805-50d03fee9a78-apiservice-cert\") pod \"metallb-operator-controller-manager-c8bf5ddcb-9s6db\" (UID: \"e75166b5-2b64-4717-9805-50d03fee9a78\") " pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.673633 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e75166b5-2b64-4717-9805-50d03fee9a78-webhook-cert\") pod \"metallb-operator-controller-manager-c8bf5ddcb-9s6db\" (UID: \"e75166b5-2b64-4717-9805-50d03fee9a78\") " pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.673670 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56l4n\" (UniqueName: \"kubernetes.io/projected/e75166b5-2b64-4717-9805-50d03fee9a78-kube-api-access-56l4n\") pod \"metallb-operator-controller-manager-c8bf5ddcb-9s6db\" (UID: \"e75166b5-2b64-4717-9805-50d03fee9a78\") " pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.770689 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b"] Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.771522 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.773483 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.773746 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-2m6zp" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.774134 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.774516 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e75166b5-2b64-4717-9805-50d03fee9a78-apiservice-cert\") pod \"metallb-operator-controller-manager-c8bf5ddcb-9s6db\" (UID: \"e75166b5-2b64-4717-9805-50d03fee9a78\") " pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.774587 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e75166b5-2b64-4717-9805-50d03fee9a78-webhook-cert\") pod \"metallb-operator-controller-manager-c8bf5ddcb-9s6db\" (UID: \"e75166b5-2b64-4717-9805-50d03fee9a78\") " pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.774635 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56l4n\" (UniqueName: \"kubernetes.io/projected/e75166b5-2b64-4717-9805-50d03fee9a78-kube-api-access-56l4n\") pod \"metallb-operator-controller-manager-c8bf5ddcb-9s6db\" (UID: \"e75166b5-2b64-4717-9805-50d03fee9a78\") " pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.781181 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e75166b5-2b64-4717-9805-50d03fee9a78-apiservice-cert\") pod \"metallb-operator-controller-manager-c8bf5ddcb-9s6db\" (UID: \"e75166b5-2b64-4717-9805-50d03fee9a78\") " pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.781628 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e75166b5-2b64-4717-9805-50d03fee9a78-webhook-cert\") pod \"metallb-operator-controller-manager-c8bf5ddcb-9s6db\" (UID: \"e75166b5-2b64-4717-9805-50d03fee9a78\") " pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.806713 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b"] Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.814673 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56l4n\" (UniqueName: \"kubernetes.io/projected/e75166b5-2b64-4717-9805-50d03fee9a78-kube-api-access-56l4n\") pod \"metallb-operator-controller-manager-c8bf5ddcb-9s6db\" (UID: \"e75166b5-2b64-4717-9805-50d03fee9a78\") " pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.859255 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.875489 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5cd2a2a-3443-4865-af85-d2da6ca9900a-webhook-cert\") pod \"metallb-operator-webhook-server-659bb64868-tcb2b\" (UID: \"e5cd2a2a-3443-4865-af85-d2da6ca9900a\") " pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.875561 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5cd2a2a-3443-4865-af85-d2da6ca9900a-apiservice-cert\") pod \"metallb-operator-webhook-server-659bb64868-tcb2b\" (UID: \"e5cd2a2a-3443-4865-af85-d2da6ca9900a\") " pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.875634 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btvgf\" (UniqueName: \"kubernetes.io/projected/e5cd2a2a-3443-4865-af85-d2da6ca9900a-kube-api-access-btvgf\") pod \"metallb-operator-webhook-server-659bb64868-tcb2b\" (UID: \"e5cd2a2a-3443-4865-af85-d2da6ca9900a\") " pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.976911 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btvgf\" (UniqueName: \"kubernetes.io/projected/e5cd2a2a-3443-4865-af85-d2da6ca9900a-kube-api-access-btvgf\") pod \"metallb-operator-webhook-server-659bb64868-tcb2b\" (UID: \"e5cd2a2a-3443-4865-af85-d2da6ca9900a\") " pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.976961 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5cd2a2a-3443-4865-af85-d2da6ca9900a-webhook-cert\") pod \"metallb-operator-webhook-server-659bb64868-tcb2b\" (UID: \"e5cd2a2a-3443-4865-af85-d2da6ca9900a\") " pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.976998 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5cd2a2a-3443-4865-af85-d2da6ca9900a-apiservice-cert\") pod \"metallb-operator-webhook-server-659bb64868-tcb2b\" (UID: \"e5cd2a2a-3443-4865-af85-d2da6ca9900a\") " pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.989112 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5cd2a2a-3443-4865-af85-d2da6ca9900a-webhook-cert\") pod \"metallb-operator-webhook-server-659bb64868-tcb2b\" (UID: \"e5cd2a2a-3443-4865-af85-d2da6ca9900a\") " pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:04 crc kubenswrapper[4770]: I1004 03:16:04.993981 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5cd2a2a-3443-4865-af85-d2da6ca9900a-apiservice-cert\") pod \"metallb-operator-webhook-server-659bb64868-tcb2b\" (UID: \"e5cd2a2a-3443-4865-af85-d2da6ca9900a\") " pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:05 crc kubenswrapper[4770]: I1004 03:16:05.001861 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btvgf\" (UniqueName: \"kubernetes.io/projected/e5cd2a2a-3443-4865-af85-d2da6ca9900a-kube-api-access-btvgf\") pod \"metallb-operator-webhook-server-659bb64868-tcb2b\" (UID: \"e5cd2a2a-3443-4865-af85-d2da6ca9900a\") " pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:05 crc kubenswrapper[4770]: I1004 03:16:05.154727 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:05 crc kubenswrapper[4770]: I1004 03:16:05.349282 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db"] Oct 04 03:16:05 crc kubenswrapper[4770]: W1004 03:16:05.379290 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode75166b5_2b64_4717_9805_50d03fee9a78.slice/crio-c85daddc8b1d112c85582f512727b644c16e65e55e27e36c3341eef54c0353e4 WatchSource:0}: Error finding container c85daddc8b1d112c85582f512727b644c16e65e55e27e36c3341eef54c0353e4: Status 404 returned error can't find the container with id c85daddc8b1d112c85582f512727b644c16e65e55e27e36c3341eef54c0353e4 Oct 04 03:16:05 crc kubenswrapper[4770]: I1004 03:16:05.568493 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" event={"ID":"e75166b5-2b64-4717-9805-50d03fee9a78","Type":"ContainerStarted","Data":"c85daddc8b1d112c85582f512727b644c16e65e55e27e36c3341eef54c0353e4"} Oct 04 03:16:05 crc kubenswrapper[4770]: I1004 03:16:05.634434 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b"] Oct 04 03:16:05 crc kubenswrapper[4770]: W1004 03:16:05.644282 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5cd2a2a_3443_4865_af85_d2da6ca9900a.slice/crio-8269164de06f9133e290b4740383af1f239fb5884c368186e49fc49b2f0c4d82 WatchSource:0}: Error finding container 8269164de06f9133e290b4740383af1f239fb5884c368186e49fc49b2f0c4d82: Status 404 returned error can't find the container with id 8269164de06f9133e290b4740383af1f239fb5884c368186e49fc49b2f0c4d82 Oct 04 03:16:05 crc kubenswrapper[4770]: I1004 03:16:05.681194 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34745ebf-06a4-4019-beb3-4fd2e2855150" path="/var/lib/kubelet/pods/34745ebf-06a4-4019-beb3-4fd2e2855150/volumes" Oct 04 03:16:06 crc kubenswrapper[4770]: I1004 03:16:06.576206 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" event={"ID":"e5cd2a2a-3443-4865-af85-d2da6ca9900a","Type":"ContainerStarted","Data":"8269164de06f9133e290b4740383af1f239fb5884c368186e49fc49b2f0c4d82"} Oct 04 03:16:09 crc kubenswrapper[4770]: I1004 03:16:09.597207 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" event={"ID":"e75166b5-2b64-4717-9805-50d03fee9a78","Type":"ContainerStarted","Data":"2ee6324ecc284dd4ed9492bafa2d51f104abc44a98206ebf126f0c9019a6bcc4"} Oct 04 03:16:09 crc kubenswrapper[4770]: I1004 03:16:09.597644 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:09 crc kubenswrapper[4770]: I1004 03:16:09.639710 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" podStartSLOduration=2.271630306 podStartE2EDuration="5.639484039s" podCreationTimestamp="2025-10-04 03:16:04 +0000 UTC" firstStartedPulling="2025-10-04 03:16:05.385581061 +0000 UTC m=+776.677590763" lastFinishedPulling="2025-10-04 03:16:08.753434774 +0000 UTC m=+780.045444496" observedRunningTime="2025-10-04 03:16:09.636514311 +0000 UTC m=+780.928524033" watchObservedRunningTime="2025-10-04 03:16:09.639484039 +0000 UTC m=+780.931493761" Oct 04 03:16:11 crc kubenswrapper[4770]: I1004 03:16:11.618628 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" event={"ID":"e5cd2a2a-3443-4865-af85-d2da6ca9900a","Type":"ContainerStarted","Data":"dd31a8604bf65f3e04276d45ad6f64fb61fef793f73e7a50d2a07130fa77a9bc"} Oct 04 03:16:11 crc kubenswrapper[4770]: I1004 03:16:11.619654 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:11 crc kubenswrapper[4770]: I1004 03:16:11.644669 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" podStartSLOduration=2.404933606 podStartE2EDuration="7.644642652s" podCreationTimestamp="2025-10-04 03:16:04 +0000 UTC" firstStartedPulling="2025-10-04 03:16:05.647826768 +0000 UTC m=+776.939836480" lastFinishedPulling="2025-10-04 03:16:10.887535814 +0000 UTC m=+782.179545526" observedRunningTime="2025-10-04 03:16:11.641651274 +0000 UTC m=+782.933661046" watchObservedRunningTime="2025-10-04 03:16:11.644642652 +0000 UTC m=+782.936652404" Oct 04 03:16:25 crc kubenswrapper[4770]: I1004 03:16:25.161892 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-659bb64868-tcb2b" Oct 04 03:16:31 crc kubenswrapper[4770]: I1004 03:16:31.795868 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:16:31 crc kubenswrapper[4770]: I1004 03:16:31.796734 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:16:31 crc kubenswrapper[4770]: I1004 03:16:31.886785 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vdrf6"] Oct 04 03:16:31 crc kubenswrapper[4770]: I1004 03:16:31.888703 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:31 crc kubenswrapper[4770]: I1004 03:16:31.894051 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vdrf6"] Oct 04 03:16:31 crc kubenswrapper[4770]: I1004 03:16:31.981725 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-catalog-content\") pod \"certified-operators-vdrf6\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:31 crc kubenswrapper[4770]: I1004 03:16:31.982085 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-utilities\") pod \"certified-operators-vdrf6\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:31 crc kubenswrapper[4770]: I1004 03:16:31.982134 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfms2\" (UniqueName: \"kubernetes.io/projected/3f678748-7332-4900-8c2e-9b9b3111136f-kube-api-access-bfms2\") pod \"certified-operators-vdrf6\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:32 crc kubenswrapper[4770]: I1004 03:16:32.083720 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfms2\" (UniqueName: \"kubernetes.io/projected/3f678748-7332-4900-8c2e-9b9b3111136f-kube-api-access-bfms2\") pod \"certified-operators-vdrf6\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:32 crc kubenswrapper[4770]: I1004 03:16:32.083807 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-catalog-content\") pod \"certified-operators-vdrf6\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:32 crc kubenswrapper[4770]: I1004 03:16:32.083828 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-utilities\") pod \"certified-operators-vdrf6\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:32 crc kubenswrapper[4770]: I1004 03:16:32.084288 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-utilities\") pod \"certified-operators-vdrf6\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:32 crc kubenswrapper[4770]: I1004 03:16:32.084459 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-catalog-content\") pod \"certified-operators-vdrf6\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:32 crc kubenswrapper[4770]: I1004 03:16:32.113875 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfms2\" (UniqueName: \"kubernetes.io/projected/3f678748-7332-4900-8c2e-9b9b3111136f-kube-api-access-bfms2\") pod \"certified-operators-vdrf6\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:32 crc kubenswrapper[4770]: I1004 03:16:32.227111 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:32 crc kubenswrapper[4770]: I1004 03:16:32.685212 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vdrf6"] Oct 04 03:16:32 crc kubenswrapper[4770]: I1004 03:16:32.772787 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrf6" event={"ID":"3f678748-7332-4900-8c2e-9b9b3111136f","Type":"ContainerStarted","Data":"4678065ae9b05a5b5cac993d12bca20244c8e426eeb4e4ed193c9961525d47bf"} Oct 04 03:16:33 crc kubenswrapper[4770]: I1004 03:16:33.794292 4770 generic.go:334] "Generic (PLEG): container finished" podID="3f678748-7332-4900-8c2e-9b9b3111136f" containerID="cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4" exitCode=0 Oct 04 03:16:33 crc kubenswrapper[4770]: I1004 03:16:33.794377 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrf6" event={"ID":"3f678748-7332-4900-8c2e-9b9b3111136f","Type":"ContainerDied","Data":"cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4"} Oct 04 03:16:35 crc kubenswrapper[4770]: I1004 03:16:35.810256 4770 generic.go:334] "Generic (PLEG): container finished" podID="3f678748-7332-4900-8c2e-9b9b3111136f" containerID="a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f" exitCode=0 Oct 04 03:16:35 crc kubenswrapper[4770]: I1004 03:16:35.810318 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrf6" event={"ID":"3f678748-7332-4900-8c2e-9b9b3111136f","Type":"ContainerDied","Data":"a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f"} Oct 04 03:16:36 crc kubenswrapper[4770]: I1004 03:16:36.826606 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrf6" event={"ID":"3f678748-7332-4900-8c2e-9b9b3111136f","Type":"ContainerStarted","Data":"802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3"} Oct 04 03:16:36 crc kubenswrapper[4770]: I1004 03:16:36.857434 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vdrf6" podStartSLOduration=3.193240667 podStartE2EDuration="5.857403507s" podCreationTimestamp="2025-10-04 03:16:31 +0000 UTC" firstStartedPulling="2025-10-04 03:16:33.798888401 +0000 UTC m=+805.090898113" lastFinishedPulling="2025-10-04 03:16:36.463051201 +0000 UTC m=+807.755060953" observedRunningTime="2025-10-04 03:16:36.852072348 +0000 UTC m=+808.144082100" watchObservedRunningTime="2025-10-04 03:16:36.857403507 +0000 UTC m=+808.149413229" Oct 04 03:16:42 crc kubenswrapper[4770]: I1004 03:16:42.227638 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:42 crc kubenswrapper[4770]: I1004 03:16:42.228214 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:42 crc kubenswrapper[4770]: I1004 03:16:42.298698 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:42 crc kubenswrapper[4770]: I1004 03:16:42.957123 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:43 crc kubenswrapper[4770]: I1004 03:16:43.034541 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vdrf6"] Oct 04 03:16:44 crc kubenswrapper[4770]: I1004 03:16:44.865177 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-c8bf5ddcb-9s6db" Oct 04 03:16:44 crc kubenswrapper[4770]: I1004 03:16:44.889596 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vdrf6" podUID="3f678748-7332-4900-8c2e-9b9b3111136f" containerName="registry-server" containerID="cri-o://802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3" gracePeriod=2 Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.325644 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.388415 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfms2\" (UniqueName: \"kubernetes.io/projected/3f678748-7332-4900-8c2e-9b9b3111136f-kube-api-access-bfms2\") pod \"3f678748-7332-4900-8c2e-9b9b3111136f\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.388495 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-utilities\") pod \"3f678748-7332-4900-8c2e-9b9b3111136f\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.388672 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-catalog-content\") pod \"3f678748-7332-4900-8c2e-9b9b3111136f\" (UID: \"3f678748-7332-4900-8c2e-9b9b3111136f\") " Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.389921 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-utilities" (OuterVolumeSpecName: "utilities") pod "3f678748-7332-4900-8c2e-9b9b3111136f" (UID: "3f678748-7332-4900-8c2e-9b9b3111136f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.396714 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f678748-7332-4900-8c2e-9b9b3111136f-kube-api-access-bfms2" (OuterVolumeSpecName: "kube-api-access-bfms2") pod "3f678748-7332-4900-8c2e-9b9b3111136f" (UID: "3f678748-7332-4900-8c2e-9b9b3111136f"). InnerVolumeSpecName "kube-api-access-bfms2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.490797 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfms2\" (UniqueName: \"kubernetes.io/projected/3f678748-7332-4900-8c2e-9b9b3111136f-kube-api-access-bfms2\") on node \"crc\" DevicePath \"\"" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.490856 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.577816 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f678748-7332-4900-8c2e-9b9b3111136f" (UID: "3f678748-7332-4900-8c2e-9b9b3111136f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.592175 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f678748-7332-4900-8c2e-9b9b3111136f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.722569 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-crfm8"] Oct 04 03:16:45 crc kubenswrapper[4770]: E1004 03:16:45.723221 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f678748-7332-4900-8c2e-9b9b3111136f" containerName="registry-server" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.723246 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f678748-7332-4900-8c2e-9b9b3111136f" containerName="registry-server" Oct 04 03:16:45 crc kubenswrapper[4770]: E1004 03:16:45.723280 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f678748-7332-4900-8c2e-9b9b3111136f" containerName="extract-content" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.723294 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f678748-7332-4900-8c2e-9b9b3111136f" containerName="extract-content" Oct 04 03:16:45 crc kubenswrapper[4770]: E1004 03:16:45.723304 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f678748-7332-4900-8c2e-9b9b3111136f" containerName="extract-utilities" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.723312 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f678748-7332-4900-8c2e-9b9b3111136f" containerName="extract-utilities" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.723603 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f678748-7332-4900-8c2e-9b9b3111136f" containerName="registry-server" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.737030 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.748157 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.748570 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.749113 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-jtclz" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.751886 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-rp264"] Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.754995 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.758678 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.763776 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-rp264"] Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.794149 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-reloader\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.794198 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvsvf\" (UniqueName: \"kubernetes.io/projected/611c30a9-d9f8-4edd-a69e-07b6eda06a0e-kube-api-access-pvsvf\") pod \"frr-k8s-webhook-server-64bf5d555-rp264\" (UID: \"611c30a9-d9f8-4edd-a69e-07b6eda06a0e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.794245 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/611c30a9-d9f8-4edd-a69e-07b6eda06a0e-cert\") pod \"frr-k8s-webhook-server-64bf5d555-rp264\" (UID: \"611c30a9-d9f8-4edd-a69e-07b6eda06a0e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.794261 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-metrics\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.794278 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-frr-conf\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.794312 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhb4l\" (UniqueName: \"kubernetes.io/projected/78a45309-f737-43ad-ab27-8a0f747536a5-kube-api-access-dhb4l\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.794337 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/78a45309-f737-43ad-ab27-8a0f747536a5-frr-startup\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.794353 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-frr-sockets\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.794372 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78a45309-f737-43ad-ab27-8a0f747536a5-metrics-certs\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.839320 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-4xpdt"] Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.840277 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4xpdt" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.842201 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.842577 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.842914 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kgq6f" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.851401 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.858723 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-5vpbh"] Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.859631 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.862471 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.882770 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-5vpbh"] Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915065 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c6436ad-4593-41ac-99e1-81e41edcaf00-cert\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915129 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhb4l\" (UniqueName: \"kubernetes.io/projected/78a45309-f737-43ad-ab27-8a0f747536a5-kube-api-access-dhb4l\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915166 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfjm8\" (UniqueName: \"kubernetes.io/projected/6c6436ad-4593-41ac-99e1-81e41edcaf00-kube-api-access-mfjm8\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915191 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/78a45309-f737-43ad-ab27-8a0f747536a5-frr-startup\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915212 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-frr-sockets\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915231 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-metrics-certs\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915253 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78a45309-f737-43ad-ab27-8a0f747536a5-metrics-certs\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: E1004 03:16:45.915490 4770 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915512 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-metallb-excludel2\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:45 crc kubenswrapper[4770]: E1004 03:16:45.915595 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/78a45309-f737-43ad-ab27-8a0f747536a5-metrics-certs podName:78a45309-f737-43ad-ab27-8a0f747536a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:16:46.415554042 +0000 UTC m=+817.707563754 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/78a45309-f737-43ad-ab27-8a0f747536a5-metrics-certs") pod "frr-k8s-crfm8" (UID: "78a45309-f737-43ad-ab27-8a0f747536a5") : secret "frr-k8s-certs-secret" not found Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915622 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-reloader\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915642 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-frr-sockets\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915650 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvsvf\" (UniqueName: \"kubernetes.io/projected/611c30a9-d9f8-4edd-a69e-07b6eda06a0e-kube-api-access-pvsvf\") pod \"frr-k8s-webhook-server-64bf5d555-rp264\" (UID: \"611c30a9-d9f8-4edd-a69e-07b6eda06a0e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915706 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfvfr\" (UniqueName: \"kubernetes.io/projected/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-kube-api-access-vfvfr\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915725 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c6436ad-4593-41ac-99e1-81e41edcaf00-metrics-certs\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915777 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915800 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/611c30a9-d9f8-4edd-a69e-07b6eda06a0e-cert\") pod \"frr-k8s-webhook-server-64bf5d555-rp264\" (UID: \"611c30a9-d9f8-4edd-a69e-07b6eda06a0e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915836 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-metrics\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915876 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-frr-conf\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.915836 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-reloader\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.916471 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-frr-conf\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.916549 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/78a45309-f737-43ad-ab27-8a0f747536a5-metrics\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.917128 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/78a45309-f737-43ad-ab27-8a0f747536a5-frr-startup\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.920857 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/611c30a9-d9f8-4edd-a69e-07b6eda06a0e-cert\") pod \"frr-k8s-webhook-server-64bf5d555-rp264\" (UID: \"611c30a9-d9f8-4edd-a69e-07b6eda06a0e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.934017 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhb4l\" (UniqueName: \"kubernetes.io/projected/78a45309-f737-43ad-ab27-8a0f747536a5-kube-api-access-dhb4l\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.939402 4770 generic.go:334] "Generic (PLEG): container finished" podID="3f678748-7332-4900-8c2e-9b9b3111136f" containerID="802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3" exitCode=0 Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.939460 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrf6" event={"ID":"3f678748-7332-4900-8c2e-9b9b3111136f","Type":"ContainerDied","Data":"802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3"} Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.939494 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vdrf6" event={"ID":"3f678748-7332-4900-8c2e-9b9b3111136f","Type":"ContainerDied","Data":"4678065ae9b05a5b5cac993d12bca20244c8e426eeb4e4ed193c9961525d47bf"} Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.939515 4770 scope.go:117] "RemoveContainer" containerID="802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.939706 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vdrf6" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.943390 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvsvf\" (UniqueName: \"kubernetes.io/projected/611c30a9-d9f8-4edd-a69e-07b6eda06a0e-kube-api-access-pvsvf\") pod \"frr-k8s-webhook-server-64bf5d555-rp264\" (UID: \"611c30a9-d9f8-4edd-a69e-07b6eda06a0e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.979422 4770 scope.go:117] "RemoveContainer" containerID="a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f" Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.984821 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vdrf6"] Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.994120 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vdrf6"] Oct 04 03:16:45 crc kubenswrapper[4770]: I1004 03:16:45.998266 4770 scope.go:117] "RemoveContainer" containerID="cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.016203 4770 scope.go:117] "RemoveContainer" containerID="802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.017661 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-metallb-excludel2\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.017716 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c6436ad-4593-41ac-99e1-81e41edcaf00-metrics-certs\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.017739 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfvfr\" (UniqueName: \"kubernetes.io/projected/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-kube-api-access-vfvfr\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.017767 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.017809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c6436ad-4593-41ac-99e1-81e41edcaf00-cert\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.017835 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfjm8\" (UniqueName: \"kubernetes.io/projected/6c6436ad-4593-41ac-99e1-81e41edcaf00-kube-api-access-mfjm8\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.017855 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-metrics-certs\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:46 crc kubenswrapper[4770]: E1004 03:16:46.018267 4770 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 03:16:46 crc kubenswrapper[4770]: E1004 03:16:46.018394 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist podName:1ee98b5e-b3c0-4c5e-a99c-332d513bc539 nodeName:}" failed. No retries permitted until 2025-10-04 03:16:46.518360836 +0000 UTC m=+817.810370548 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist") pod "speaker-4xpdt" (UID: "1ee98b5e-b3c0-4c5e-a99c-332d513bc539") : secret "metallb-memberlist" not found Oct 04 03:16:46 crc kubenswrapper[4770]: E1004 03:16:46.018866 4770 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 04 03:16:46 crc kubenswrapper[4770]: E1004 03:16:46.018898 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c6436ad-4593-41ac-99e1-81e41edcaf00-metrics-certs podName:6c6436ad-4593-41ac-99e1-81e41edcaf00 nodeName:}" failed. No retries permitted until 2025-10-04 03:16:46.5188878 +0000 UTC m=+817.810897522 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c6436ad-4593-41ac-99e1-81e41edcaf00-metrics-certs") pod "controller-68d546b9d8-5vpbh" (UID: "6c6436ad-4593-41ac-99e1-81e41edcaf00") : secret "controller-certs-secret" not found Oct 04 03:16:46 crc kubenswrapper[4770]: E1004 03:16:46.019280 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3\": container with ID starting with 802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3 not found: ID does not exist" containerID="802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.019323 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3"} err="failed to get container status \"802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3\": rpc error: code = NotFound desc = could not find container \"802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3\": container with ID starting with 802060f44db433df21e24f1a5fefc6e7f0b2bcd062b504f26ccea01921be35f3 not found: ID does not exist" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.019353 4770 scope.go:117] "RemoveContainer" containerID="a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.019356 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-metallb-excludel2\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:46 crc kubenswrapper[4770]: E1004 03:16:46.023138 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f\": container with ID starting with a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f not found: ID does not exist" containerID="a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.023175 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f"} err="failed to get container status \"a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f\": rpc error: code = NotFound desc = could not find container \"a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f\": container with ID starting with a44613acb7d097ec95dcc252d94d6ec946bfc9c9d59cd3ba7e3c8de1dced482f not found: ID does not exist" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.023196 4770 scope.go:117] "RemoveContainer" containerID="cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.023499 4770 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.023706 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-metrics-certs\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:46 crc kubenswrapper[4770]: E1004 03:16:46.027139 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4\": container with ID starting with cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4 not found: ID does not exist" containerID="cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.027181 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4"} err="failed to get container status \"cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4\": rpc error: code = NotFound desc = could not find container \"cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4\": container with ID starting with cfc4889cf5685c4239aafcd17c068690b91911fc67653d75c905b479b4ffa0b4 not found: ID does not exist" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.032161 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c6436ad-4593-41ac-99e1-81e41edcaf00-cert\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.040579 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfvfr\" (UniqueName: \"kubernetes.io/projected/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-kube-api-access-vfvfr\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.043947 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfjm8\" (UniqueName: \"kubernetes.io/projected/6c6436ad-4593-41ac-99e1-81e41edcaf00-kube-api-access-mfjm8\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.069898 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.423967 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78a45309-f737-43ad-ab27-8a0f747536a5-metrics-certs\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.436431 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78a45309-f737-43ad-ab27-8a0f747536a5-metrics-certs\") pod \"frr-k8s-crfm8\" (UID: \"78a45309-f737-43ad-ab27-8a0f747536a5\") " pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.525680 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-rp264"] Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.526898 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c6436ad-4593-41ac-99e1-81e41edcaf00-metrics-certs\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.527184 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:46 crc kubenswrapper[4770]: E1004 03:16:46.527450 4770 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 03:16:46 crc kubenswrapper[4770]: E1004 03:16:46.527576 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist podName:1ee98b5e-b3c0-4c5e-a99c-332d513bc539 nodeName:}" failed. No retries permitted until 2025-10-04 03:16:47.52754017 +0000 UTC m=+818.819549922 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist") pod "speaker-4xpdt" (UID: "1ee98b5e-b3c0-4c5e-a99c-332d513bc539") : secret "metallb-memberlist" not found Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.532779 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c6436ad-4593-41ac-99e1-81e41edcaf00-metrics-certs\") pod \"controller-68d546b9d8-5vpbh\" (UID: \"6c6436ad-4593-41ac-99e1-81e41edcaf00\") " pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.660910 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-crfm8" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.773411 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.948402 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerStarted","Data":"714e38add2b8a0e8f162ae075aa41941945df1cd44f3acf8c576365a4a2dcd39"} Oct 04 03:16:46 crc kubenswrapper[4770]: I1004 03:16:46.950960 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" event={"ID":"611c30a9-d9f8-4edd-a69e-07b6eda06a0e","Type":"ContainerStarted","Data":"29b61a54949eb3b9ae81a4df866772028cc83b8abb9f1e32987e7b70aea28cd4"} Oct 04 03:16:47 crc kubenswrapper[4770]: I1004 03:16:47.339810 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-5vpbh"] Oct 04 03:16:47 crc kubenswrapper[4770]: W1004 03:16:47.348345 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c6436ad_4593_41ac_99e1_81e41edcaf00.slice/crio-22d38d9b59583bc3d355ac6327ca8b9d13e44e3ecf1cd3511215ab8f1b829f99 WatchSource:0}: Error finding container 22d38d9b59583bc3d355ac6327ca8b9d13e44e3ecf1cd3511215ab8f1b829f99: Status 404 returned error can't find the container with id 22d38d9b59583bc3d355ac6327ca8b9d13e44e3ecf1cd3511215ab8f1b829f99 Oct 04 03:16:47 crc kubenswrapper[4770]: I1004 03:16:47.547184 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:47 crc kubenswrapper[4770]: E1004 03:16:47.547359 4770 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 03:16:47 crc kubenswrapper[4770]: E1004 03:16:47.547424 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist podName:1ee98b5e-b3c0-4c5e-a99c-332d513bc539 nodeName:}" failed. No retries permitted until 2025-10-04 03:16:49.547408728 +0000 UTC m=+820.839418440 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist") pod "speaker-4xpdt" (UID: "1ee98b5e-b3c0-4c5e-a99c-332d513bc539") : secret "metallb-memberlist" not found Oct 04 03:16:47 crc kubenswrapper[4770]: I1004 03:16:47.686789 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f678748-7332-4900-8c2e-9b9b3111136f" path="/var/lib/kubelet/pods/3f678748-7332-4900-8c2e-9b9b3111136f/volumes" Oct 04 03:16:47 crc kubenswrapper[4770]: I1004 03:16:47.964341 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-5vpbh" event={"ID":"6c6436ad-4593-41ac-99e1-81e41edcaf00","Type":"ContainerStarted","Data":"6f45f059bbde5584939412dba1bdf9da2f3699622d92ec827432c4c61967fdbe"} Oct 04 03:16:47 crc kubenswrapper[4770]: I1004 03:16:47.964413 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-5vpbh" event={"ID":"6c6436ad-4593-41ac-99e1-81e41edcaf00","Type":"ContainerStarted","Data":"0b458ea0d1f8b398e1dabce15aacb2dd4f01b14d4d235b764ddef73213156baa"} Oct 04 03:16:47 crc kubenswrapper[4770]: I1004 03:16:47.964431 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-5vpbh" event={"ID":"6c6436ad-4593-41ac-99e1-81e41edcaf00","Type":"ContainerStarted","Data":"22d38d9b59583bc3d355ac6327ca8b9d13e44e3ecf1cd3511215ab8f1b829f99"} Oct 04 03:16:47 crc kubenswrapper[4770]: I1004 03:16:47.965018 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:16:47 crc kubenswrapper[4770]: I1004 03:16:47.986949 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-5vpbh" podStartSLOduration=2.9869304740000002 podStartE2EDuration="2.986930474s" podCreationTimestamp="2025-10-04 03:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:16:47.983601298 +0000 UTC m=+819.275611010" watchObservedRunningTime="2025-10-04 03:16:47.986930474 +0000 UTC m=+819.278940186" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.159288 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gdswj"] Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.160411 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.218418 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gdswj"] Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.258106 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-utilities\") pod \"community-operators-gdswj\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.258208 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-catalog-content\") pod \"community-operators-gdswj\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.258239 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm2ck\" (UniqueName: \"kubernetes.io/projected/fbb6aa81-4425-4666-9d9e-7053b90b6981-kube-api-access-dm2ck\") pod \"community-operators-gdswj\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.359269 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-utilities\") pod \"community-operators-gdswj\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.359369 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-catalog-content\") pod \"community-operators-gdswj\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.359396 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm2ck\" (UniqueName: \"kubernetes.io/projected/fbb6aa81-4425-4666-9d9e-7053b90b6981-kube-api-access-dm2ck\") pod \"community-operators-gdswj\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.359790 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-utilities\") pod \"community-operators-gdswj\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.359999 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-catalog-content\") pod \"community-operators-gdswj\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.380748 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm2ck\" (UniqueName: \"kubernetes.io/projected/fbb6aa81-4425-4666-9d9e-7053b90b6981-kube-api-access-dm2ck\") pod \"community-operators-gdswj\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.475083 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.772317 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gdswj"] Oct 04 03:16:48 crc kubenswrapper[4770]: W1004 03:16:48.812491 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbb6aa81_4425_4666_9d9e_7053b90b6981.slice/crio-4b687c64c0673f83c4e46f5f413f78a6937195d91ffa19de9e7d07cabb3c237a WatchSource:0}: Error finding container 4b687c64c0673f83c4e46f5f413f78a6937195d91ffa19de9e7d07cabb3c237a: Status 404 returned error can't find the container with id 4b687c64c0673f83c4e46f5f413f78a6937195d91ffa19de9e7d07cabb3c237a Oct 04 03:16:48 crc kubenswrapper[4770]: I1004 03:16:48.974474 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdswj" event={"ID":"fbb6aa81-4425-4666-9d9e-7053b90b6981","Type":"ContainerStarted","Data":"4b687c64c0673f83c4e46f5f413f78a6937195d91ffa19de9e7d07cabb3c237a"} Oct 04 03:16:49 crc kubenswrapper[4770]: I1004 03:16:49.579688 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:49 crc kubenswrapper[4770]: I1004 03:16:49.605056 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1ee98b5e-b3c0-4c5e-a99c-332d513bc539-memberlist\") pod \"speaker-4xpdt\" (UID: \"1ee98b5e-b3c0-4c5e-a99c-332d513bc539\") " pod="metallb-system/speaker-4xpdt" Oct 04 03:16:49 crc kubenswrapper[4770]: I1004 03:16:49.753135 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4xpdt" Oct 04 03:16:49 crc kubenswrapper[4770]: W1004 03:16:49.776899 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ee98b5e_b3c0_4c5e_a99c_332d513bc539.slice/crio-a3cba1bc3186c195c37cccf732a8c58ee980fc34fa2c91d46dfb627593fa4ab3 WatchSource:0}: Error finding container a3cba1bc3186c195c37cccf732a8c58ee980fc34fa2c91d46dfb627593fa4ab3: Status 404 returned error can't find the container with id a3cba1bc3186c195c37cccf732a8c58ee980fc34fa2c91d46dfb627593fa4ab3 Oct 04 03:16:49 crc kubenswrapper[4770]: I1004 03:16:49.983844 4770 generic.go:334] "Generic (PLEG): container finished" podID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerID="df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7" exitCode=0 Oct 04 03:16:49 crc kubenswrapper[4770]: I1004 03:16:49.983971 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdswj" event={"ID":"fbb6aa81-4425-4666-9d9e-7053b90b6981","Type":"ContainerDied","Data":"df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7"} Oct 04 03:16:49 crc kubenswrapper[4770]: I1004 03:16:49.987395 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4xpdt" event={"ID":"1ee98b5e-b3c0-4c5e-a99c-332d513bc539","Type":"ContainerStarted","Data":"a3cba1bc3186c195c37cccf732a8c58ee980fc34fa2c91d46dfb627593fa4ab3"} Oct 04 03:16:51 crc kubenswrapper[4770]: I1004 03:16:50.999360 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4xpdt" event={"ID":"1ee98b5e-b3c0-4c5e-a99c-332d513bc539","Type":"ContainerStarted","Data":"ead8c8b76befc02d9451ac68c0cabaa5261df1732769a453d7bed49a73426778"} Oct 04 03:16:51 crc kubenswrapper[4770]: I1004 03:16:50.999854 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4xpdt" event={"ID":"1ee98b5e-b3c0-4c5e-a99c-332d513bc539","Type":"ContainerStarted","Data":"837b73619d921ae84121bf7a4177f0b1486025b8a3ac49d6b5f9d6d578875d2d"} Oct 04 03:16:51 crc kubenswrapper[4770]: I1004 03:16:50.999878 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-4xpdt" Oct 04 03:16:51 crc kubenswrapper[4770]: I1004 03:16:51.023306 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-4xpdt" podStartSLOduration=6.023281102 podStartE2EDuration="6.023281102s" podCreationTimestamp="2025-10-04 03:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:16:51.020561711 +0000 UTC m=+822.312571443" watchObservedRunningTime="2025-10-04 03:16:51.023281102 +0000 UTC m=+822.315290804" Oct 04 03:16:52 crc kubenswrapper[4770]: I1004 03:16:52.018614 4770 generic.go:334] "Generic (PLEG): container finished" podID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerID="4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36" exitCode=0 Oct 04 03:16:52 crc kubenswrapper[4770]: I1004 03:16:52.018742 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdswj" event={"ID":"fbb6aa81-4425-4666-9d9e-7053b90b6981","Type":"ContainerDied","Data":"4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36"} Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.541673 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cgjz2"] Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.544111 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cgjz2"] Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.545029 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.651869 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-427z6\" (UniqueName: \"kubernetes.io/projected/04ca1e08-a8cb-42e4-ac54-30f87452aacf-kube-api-access-427z6\") pod \"redhat-marketplace-cgjz2\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.651970 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-utilities\") pod \"redhat-marketplace-cgjz2\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.652035 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-catalog-content\") pod \"redhat-marketplace-cgjz2\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.753423 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-utilities\") pod \"redhat-marketplace-cgjz2\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.753524 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-catalog-content\") pod \"redhat-marketplace-cgjz2\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.753665 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-427z6\" (UniqueName: \"kubernetes.io/projected/04ca1e08-a8cb-42e4-ac54-30f87452aacf-kube-api-access-427z6\") pod \"redhat-marketplace-cgjz2\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.754128 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-utilities\") pod \"redhat-marketplace-cgjz2\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.754584 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-catalog-content\") pod \"redhat-marketplace-cgjz2\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.800300 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-427z6\" (UniqueName: \"kubernetes.io/projected/04ca1e08-a8cb-42e4-ac54-30f87452aacf-kube-api-access-427z6\") pod \"redhat-marketplace-cgjz2\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:53 crc kubenswrapper[4770]: I1004 03:16:53.878944 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:16:54 crc kubenswrapper[4770]: I1004 03:16:54.696716 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cgjz2"] Oct 04 03:16:54 crc kubenswrapper[4770]: W1004 03:16:54.707153 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04ca1e08_a8cb_42e4_ac54_30f87452aacf.slice/crio-5e26d9592314ad87288a001b65965103dde18a17b903fcf7e6b4f9d9ab4f08b0 WatchSource:0}: Error finding container 5e26d9592314ad87288a001b65965103dde18a17b903fcf7e6b4f9d9ab4f08b0: Status 404 returned error can't find the container with id 5e26d9592314ad87288a001b65965103dde18a17b903fcf7e6b4f9d9ab4f08b0 Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.044853 4770 generic.go:334] "Generic (PLEG): container finished" podID="78a45309-f737-43ad-ab27-8a0f747536a5" containerID="6035a3f4bb73a42a78904253e9fdc6c7dd9e0f034fbe91d4ea54bf6563195599" exitCode=0 Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.044923 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerDied","Data":"6035a3f4bb73a42a78904253e9fdc6c7dd9e0f034fbe91d4ea54bf6563195599"} Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.049073 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" event={"ID":"611c30a9-d9f8-4edd-a69e-07b6eda06a0e","Type":"ContainerStarted","Data":"90c9255f4b2549029a269c8ec2ae97ae257121db052c9f57dde155bb75dae10b"} Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.049202 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.054846 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdswj" event={"ID":"fbb6aa81-4425-4666-9d9e-7053b90b6981","Type":"ContainerStarted","Data":"7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3"} Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.061588 4770 generic.go:334] "Generic (PLEG): container finished" podID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerID="155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd" exitCode=0 Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.061638 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cgjz2" event={"ID":"04ca1e08-a8cb-42e4-ac54-30f87452aacf","Type":"ContainerDied","Data":"155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd"} Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.061669 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cgjz2" event={"ID":"04ca1e08-a8cb-42e4-ac54-30f87452aacf","Type":"ContainerStarted","Data":"5e26d9592314ad87288a001b65965103dde18a17b903fcf7e6b4f9d9ab4f08b0"} Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.136372 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gdswj" podStartSLOduration=2.5001011909999997 podStartE2EDuration="7.136348882s" podCreationTimestamp="2025-10-04 03:16:48 +0000 UTC" firstStartedPulling="2025-10-04 03:16:49.985167607 +0000 UTC m=+821.277177319" lastFinishedPulling="2025-10-04 03:16:54.621415288 +0000 UTC m=+825.913425010" observedRunningTime="2025-10-04 03:16:55.133448367 +0000 UTC m=+826.425458099" watchObservedRunningTime="2025-10-04 03:16:55.136348882 +0000 UTC m=+826.428358594" Oct 04 03:16:55 crc kubenswrapper[4770]: I1004 03:16:55.159323 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" podStartSLOduration=2.402237817 podStartE2EDuration="10.159296692s" podCreationTimestamp="2025-10-04 03:16:45 +0000 UTC" firstStartedPulling="2025-10-04 03:16:46.549233116 +0000 UTC m=+817.841242868" lastFinishedPulling="2025-10-04 03:16:54.306292021 +0000 UTC m=+825.598301743" observedRunningTime="2025-10-04 03:16:55.1515951 +0000 UTC m=+826.443604822" watchObservedRunningTime="2025-10-04 03:16:55.159296692 +0000 UTC m=+826.451306424" Oct 04 03:16:56 crc kubenswrapper[4770]: I1004 03:16:56.083655 4770 generic.go:334] "Generic (PLEG): container finished" podID="78a45309-f737-43ad-ab27-8a0f747536a5" containerID="ec8772b47cec6870b11bd353d40c17b976e907bf3039b48b801f3a6406ee6c99" exitCode=0 Oct 04 03:16:56 crc kubenswrapper[4770]: I1004 03:16:56.085539 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerDied","Data":"ec8772b47cec6870b11bd353d40c17b976e907bf3039b48b801f3a6406ee6c99"} Oct 04 03:16:57 crc kubenswrapper[4770]: I1004 03:16:57.096593 4770 generic.go:334] "Generic (PLEG): container finished" podID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerID="ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9" exitCode=0 Oct 04 03:16:57 crc kubenswrapper[4770]: I1004 03:16:57.096747 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cgjz2" event={"ID":"04ca1e08-a8cb-42e4-ac54-30f87452aacf","Type":"ContainerDied","Data":"ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9"} Oct 04 03:16:57 crc kubenswrapper[4770]: I1004 03:16:57.104335 4770 generic.go:334] "Generic (PLEG): container finished" podID="78a45309-f737-43ad-ab27-8a0f747536a5" containerID="cef318dc90bd4e920e70e42d0194b41da960e7f5ea1de9d289b06e7e2372071f" exitCode=0 Oct 04 03:16:57 crc kubenswrapper[4770]: I1004 03:16:57.104411 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerDied","Data":"cef318dc90bd4e920e70e42d0194b41da960e7f5ea1de9d289b06e7e2372071f"} Oct 04 03:16:58 crc kubenswrapper[4770]: I1004 03:16:58.116947 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cgjz2" event={"ID":"04ca1e08-a8cb-42e4-ac54-30f87452aacf","Type":"ContainerStarted","Data":"40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3"} Oct 04 03:16:58 crc kubenswrapper[4770]: I1004 03:16:58.126722 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerStarted","Data":"1bdd7a4017ca74951c0a8dbc52f8f5cf13566cfb1128f1016ea42a8210cb7170"} Oct 04 03:16:58 crc kubenswrapper[4770]: I1004 03:16:58.126780 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerStarted","Data":"c1704aa93c9f2165d1aab6dbe4c5a6100eb0531b5c16397558c10a34c0284a6f"} Oct 04 03:16:58 crc kubenswrapper[4770]: I1004 03:16:58.126795 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerStarted","Data":"886856d2f8beedfecc6a4298a8e9785778154963fc8822981dd2cce601ea0f8a"} Oct 04 03:16:58 crc kubenswrapper[4770]: I1004 03:16:58.126810 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerStarted","Data":"652a13825ba9fdd361bbfcd828a8fe2ba47002226263ec40173dc87d409719ca"} Oct 04 03:16:58 crc kubenswrapper[4770]: I1004 03:16:58.475783 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:58 crc kubenswrapper[4770]: I1004 03:16:58.475860 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:58 crc kubenswrapper[4770]: I1004 03:16:58.538524 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:58 crc kubenswrapper[4770]: I1004 03:16:58.565686 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cgjz2" podStartSLOduration=3.114169024 podStartE2EDuration="5.565660211s" podCreationTimestamp="2025-10-04 03:16:53 +0000 UTC" firstStartedPulling="2025-10-04 03:16:55.077385643 +0000 UTC m=+826.369395385" lastFinishedPulling="2025-10-04 03:16:57.52887683 +0000 UTC m=+828.820886572" observedRunningTime="2025-10-04 03:16:58.151706162 +0000 UTC m=+829.443715924" watchObservedRunningTime="2025-10-04 03:16:58.565660211 +0000 UTC m=+829.857669923" Oct 04 03:16:59 crc kubenswrapper[4770]: I1004 03:16:59.144262 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerStarted","Data":"c1d34a98fadeaaf0233061c5b4ecdc2c7fc319d5d06a97694b63daea9e19b62d"} Oct 04 03:16:59 crc kubenswrapper[4770]: I1004 03:16:59.145380 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-crfm8" event={"ID":"78a45309-f737-43ad-ab27-8a0f747536a5","Type":"ContainerStarted","Data":"e0f811a31d812a62e84b2e8ddce6d8d60a965e8d832cd2f0d878ca8abcded503"} Oct 04 03:16:59 crc kubenswrapper[4770]: I1004 03:16:59.180277 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-crfm8" podStartSLOduration=6.693092712 podStartE2EDuration="14.180237917s" podCreationTimestamp="2025-10-04 03:16:45 +0000 UTC" firstStartedPulling="2025-10-04 03:16:46.847959617 +0000 UTC m=+818.139969369" lastFinishedPulling="2025-10-04 03:16:54.335104852 +0000 UTC m=+825.627114574" observedRunningTime="2025-10-04 03:16:59.179366554 +0000 UTC m=+830.471376266" watchObservedRunningTime="2025-10-04 03:16:59.180237917 +0000 UTC m=+830.472247629" Oct 04 03:16:59 crc kubenswrapper[4770]: I1004 03:16:59.225717 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:16:59 crc kubenswrapper[4770]: I1004 03:16:59.759638 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-4xpdt" Oct 04 03:17:00 crc kubenswrapper[4770]: I1004 03:17:00.150093 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-crfm8" Oct 04 03:17:00 crc kubenswrapper[4770]: I1004 03:17:00.713396 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gdswj"] Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.155983 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gdswj" podUID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerName="registry-server" containerID="cri-o://7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3" gracePeriod=2 Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.585629 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj"] Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.587321 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.590831 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.601947 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj"] Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.661335 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-crfm8" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.668829 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.668908 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd8h5\" (UniqueName: \"kubernetes.io/projected/714e499d-6a91-4a5a-b96f-4a56413bc7eb-kube-api-access-zd8h5\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.668945 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.714316 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-crfm8" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.770725 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.770856 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd8h5\" (UniqueName: \"kubernetes.io/projected/714e499d-6a91-4a5a-b96f-4a56413bc7eb-kube-api-access-zd8h5\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.770907 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.772288 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.772707 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.794230 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd8h5\" (UniqueName: \"kubernetes.io/projected/714e499d-6a91-4a5a-b96f-4a56413bc7eb-kube-api-access-zd8h5\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.796243 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.796315 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:17:01 crc kubenswrapper[4770]: I1004 03:17:01.912942 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.047418 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.163974 4770 generic.go:334] "Generic (PLEG): container finished" podID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerID="7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3" exitCode=0 Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.164046 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdswj" event={"ID":"fbb6aa81-4425-4666-9d9e-7053b90b6981","Type":"ContainerDied","Data":"7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3"} Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.164097 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdswj" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.164124 4770 scope.go:117] "RemoveContainer" containerID="7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.164109 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdswj" event={"ID":"fbb6aa81-4425-4666-9d9e-7053b90b6981","Type":"ContainerDied","Data":"4b687c64c0673f83c4e46f5f413f78a6937195d91ffa19de9e7d07cabb3c237a"} Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.183317 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm2ck\" (UniqueName: \"kubernetes.io/projected/fbb6aa81-4425-4666-9d9e-7053b90b6981-kube-api-access-dm2ck\") pod \"fbb6aa81-4425-4666-9d9e-7053b90b6981\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.183474 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-utilities\") pod \"fbb6aa81-4425-4666-9d9e-7053b90b6981\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.183620 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-catalog-content\") pod \"fbb6aa81-4425-4666-9d9e-7053b90b6981\" (UID: \"fbb6aa81-4425-4666-9d9e-7053b90b6981\") " Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.185379 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj"] Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.188872 4770 scope.go:117] "RemoveContainer" containerID="4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.189149 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-utilities" (OuterVolumeSpecName: "utilities") pod "fbb6aa81-4425-4666-9d9e-7053b90b6981" (UID: "fbb6aa81-4425-4666-9d9e-7053b90b6981"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.192543 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbb6aa81-4425-4666-9d9e-7053b90b6981-kube-api-access-dm2ck" (OuterVolumeSpecName: "kube-api-access-dm2ck") pod "fbb6aa81-4425-4666-9d9e-7053b90b6981" (UID: "fbb6aa81-4425-4666-9d9e-7053b90b6981"). InnerVolumeSpecName "kube-api-access-dm2ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.239288 4770 scope.go:117] "RemoveContainer" containerID="df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.251833 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fbb6aa81-4425-4666-9d9e-7053b90b6981" (UID: "fbb6aa81-4425-4666-9d9e-7053b90b6981"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.270423 4770 scope.go:117] "RemoveContainer" containerID="7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3" Oct 04 03:17:02 crc kubenswrapper[4770]: E1004 03:17:02.270896 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3\": container with ID starting with 7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3 not found: ID does not exist" containerID="7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.270963 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3"} err="failed to get container status \"7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3\": rpc error: code = NotFound desc = could not find container \"7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3\": container with ID starting with 7e463662bfdc25c8fcf131e34652146067468e7250b31b7d52e5998c88e348b3 not found: ID does not exist" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.271019 4770 scope.go:117] "RemoveContainer" containerID="4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36" Oct 04 03:17:02 crc kubenswrapper[4770]: E1004 03:17:02.271429 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36\": container with ID starting with 4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36 not found: ID does not exist" containerID="4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.271496 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36"} err="failed to get container status \"4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36\": rpc error: code = NotFound desc = could not find container \"4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36\": container with ID starting with 4f54b440758279d5342beb16691c9f43b35ef75e9ee354ac889dbf16df96bd36 not found: ID does not exist" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.271771 4770 scope.go:117] "RemoveContainer" containerID="df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7" Oct 04 03:17:02 crc kubenswrapper[4770]: E1004 03:17:02.272332 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7\": container with ID starting with df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7 not found: ID does not exist" containerID="df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.272376 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7"} err="failed to get container status \"df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7\": rpc error: code = NotFound desc = could not find container \"df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7\": container with ID starting with df4d0fe5204a5eab9c0f702d090f5970909876f49f49c0af04d76cc8640917a7 not found: ID does not exist" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.285228 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm2ck\" (UniqueName: \"kubernetes.io/projected/fbb6aa81-4425-4666-9d9e-7053b90b6981-kube-api-access-dm2ck\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.285277 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.285291 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbb6aa81-4425-4666-9d9e-7053b90b6981-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.544178 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gdswj"] Oct 04 03:17:02 crc kubenswrapper[4770]: I1004 03:17:02.550691 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gdswj"] Oct 04 03:17:03 crc kubenswrapper[4770]: I1004 03:17:03.176509 4770 generic.go:334] "Generic (PLEG): container finished" podID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerID="5b8f4f683377d985e111926b96022c1be52e1169bf74645995ad657aabc1f8c7" exitCode=0 Oct 04 03:17:03 crc kubenswrapper[4770]: I1004 03:17:03.176657 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" event={"ID":"714e499d-6a91-4a5a-b96f-4a56413bc7eb","Type":"ContainerDied","Data":"5b8f4f683377d985e111926b96022c1be52e1169bf74645995ad657aabc1f8c7"} Oct 04 03:17:03 crc kubenswrapper[4770]: I1004 03:17:03.176759 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" event={"ID":"714e499d-6a91-4a5a-b96f-4a56413bc7eb","Type":"ContainerStarted","Data":"1754017c8c5bfe3395170bea19e950f41242e0defda056c14f8305804749e903"} Oct 04 03:17:03 crc kubenswrapper[4770]: I1004 03:17:03.681894 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbb6aa81-4425-4666-9d9e-7053b90b6981" path="/var/lib/kubelet/pods/fbb6aa81-4425-4666-9d9e-7053b90b6981/volumes" Oct 04 03:17:03 crc kubenswrapper[4770]: I1004 03:17:03.879495 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:17:03 crc kubenswrapper[4770]: I1004 03:17:03.879546 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:17:03 crc kubenswrapper[4770]: I1004 03:17:03.927730 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:17:04 crc kubenswrapper[4770]: I1004 03:17:04.260759 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:17:06 crc kubenswrapper[4770]: I1004 03:17:06.075543 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rp264" Oct 04 03:17:06 crc kubenswrapper[4770]: I1004 03:17:06.778690 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-5vpbh" Oct 04 03:17:07 crc kubenswrapper[4770]: I1004 03:17:07.211284 4770 generic.go:334] "Generic (PLEG): container finished" podID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerID="cf1225961e8a691993442645278d6439e80d0d15123d22e4b7a71c57cd82765d" exitCode=0 Oct 04 03:17:07 crc kubenswrapper[4770]: I1004 03:17:07.211402 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" event={"ID":"714e499d-6a91-4a5a-b96f-4a56413bc7eb","Type":"ContainerDied","Data":"cf1225961e8a691993442645278d6439e80d0d15123d22e4b7a71c57cd82765d"} Oct 04 03:17:08 crc kubenswrapper[4770]: I1004 03:17:08.223493 4770 generic.go:334] "Generic (PLEG): container finished" podID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerID="49b0eeddc391da19c77d1b145152eeb687edc50e2bf27131a84361dc91599369" exitCode=0 Oct 04 03:17:08 crc kubenswrapper[4770]: I1004 03:17:08.223590 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" event={"ID":"714e499d-6a91-4a5a-b96f-4a56413bc7eb","Type":"ContainerDied","Data":"49b0eeddc391da19c77d1b145152eeb687edc50e2bf27131a84361dc91599369"} Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.562959 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.719810 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cgjz2"] Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.720197 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cgjz2" podUID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerName="registry-server" containerID="cri-o://40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3" gracePeriod=2 Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.725582 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-bundle\") pod \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.725715 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-util\") pod \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.725782 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd8h5\" (UniqueName: \"kubernetes.io/projected/714e499d-6a91-4a5a-b96f-4a56413bc7eb-kube-api-access-zd8h5\") pod \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\" (UID: \"714e499d-6a91-4a5a-b96f-4a56413bc7eb\") " Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.731108 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/714e499d-6a91-4a5a-b96f-4a56413bc7eb-kube-api-access-zd8h5" (OuterVolumeSpecName: "kube-api-access-zd8h5") pod "714e499d-6a91-4a5a-b96f-4a56413bc7eb" (UID: "714e499d-6a91-4a5a-b96f-4a56413bc7eb"). InnerVolumeSpecName "kube-api-access-zd8h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.733080 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-bundle" (OuterVolumeSpecName: "bundle") pod "714e499d-6a91-4a5a-b96f-4a56413bc7eb" (UID: "714e499d-6a91-4a5a-b96f-4a56413bc7eb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.755574 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-util" (OuterVolumeSpecName: "util") pod "714e499d-6a91-4a5a-b96f-4a56413bc7eb" (UID: "714e499d-6a91-4a5a-b96f-4a56413bc7eb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.828042 4770 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.828132 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd8h5\" (UniqueName: \"kubernetes.io/projected/714e499d-6a91-4a5a-b96f-4a56413bc7eb-kube-api-access-zd8h5\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:09 crc kubenswrapper[4770]: I1004 03:17:09.828154 4770 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/714e499d-6a91-4a5a-b96f-4a56413bc7eb-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.177549 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.250059 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" event={"ID":"714e499d-6a91-4a5a-b96f-4a56413bc7eb","Type":"ContainerDied","Data":"1754017c8c5bfe3395170bea19e950f41242e0defda056c14f8305804749e903"} Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.250153 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1754017c8c5bfe3395170bea19e950f41242e0defda056c14f8305804749e903" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.250115 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.256248 4770 generic.go:334] "Generic (PLEG): container finished" podID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerID="40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3" exitCode=0 Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.256350 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cgjz2" event={"ID":"04ca1e08-a8cb-42e4-ac54-30f87452aacf","Type":"ContainerDied","Data":"40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3"} Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.256414 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cgjz2" event={"ID":"04ca1e08-a8cb-42e4-ac54-30f87452aacf","Type":"ContainerDied","Data":"5e26d9592314ad87288a001b65965103dde18a17b903fcf7e6b4f9d9ab4f08b0"} Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.256428 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cgjz2" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.256451 4770 scope.go:117] "RemoveContainer" containerID="40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.278990 4770 scope.go:117] "RemoveContainer" containerID="ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.305992 4770 scope.go:117] "RemoveContainer" containerID="155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.321907 4770 scope.go:117] "RemoveContainer" containerID="40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3" Oct 04 03:17:10 crc kubenswrapper[4770]: E1004 03:17:10.322642 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3\": container with ID starting with 40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3 not found: ID does not exist" containerID="40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.322703 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3"} err="failed to get container status \"40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3\": rpc error: code = NotFound desc = could not find container \"40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3\": container with ID starting with 40fc6bf02721e185120c1b8bd90e22aecea7aeaad8fdb5968e2d1eb21fd080c3 not found: ID does not exist" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.322740 4770 scope.go:117] "RemoveContainer" containerID="ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9" Oct 04 03:17:10 crc kubenswrapper[4770]: E1004 03:17:10.323682 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9\": container with ID starting with ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9 not found: ID does not exist" containerID="ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.323733 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9"} err="failed to get container status \"ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9\": rpc error: code = NotFound desc = could not find container \"ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9\": container with ID starting with ccbc07b3646865dc81296d2d2cb4814845c4dd0cd227894c1d8c1c9abed243c9 not found: ID does not exist" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.323775 4770 scope.go:117] "RemoveContainer" containerID="155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd" Oct 04 03:17:10 crc kubenswrapper[4770]: E1004 03:17:10.324182 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd\": container with ID starting with 155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd not found: ID does not exist" containerID="155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.324218 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd"} err="failed to get container status \"155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd\": rpc error: code = NotFound desc = could not find container \"155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd\": container with ID starting with 155f49b2b6f53100eb5cb45994cc3f9a8a9feea9bb35995762b124bb6cbd3bbd not found: ID does not exist" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.334646 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-427z6\" (UniqueName: \"kubernetes.io/projected/04ca1e08-a8cb-42e4-ac54-30f87452aacf-kube-api-access-427z6\") pod \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.334739 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-utilities\") pod \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.334923 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-catalog-content\") pod \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\" (UID: \"04ca1e08-a8cb-42e4-ac54-30f87452aacf\") " Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.335793 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-utilities" (OuterVolumeSpecName: "utilities") pod "04ca1e08-a8cb-42e4-ac54-30f87452aacf" (UID: "04ca1e08-a8cb-42e4-ac54-30f87452aacf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.338641 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04ca1e08-a8cb-42e4-ac54-30f87452aacf-kube-api-access-427z6" (OuterVolumeSpecName: "kube-api-access-427z6") pod "04ca1e08-a8cb-42e4-ac54-30f87452aacf" (UID: "04ca1e08-a8cb-42e4-ac54-30f87452aacf"). InnerVolumeSpecName "kube-api-access-427z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.359584 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04ca1e08-a8cb-42e4-ac54-30f87452aacf" (UID: "04ca1e08-a8cb-42e4-ac54-30f87452aacf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.436994 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.437082 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-427z6\" (UniqueName: \"kubernetes.io/projected/04ca1e08-a8cb-42e4-ac54-30f87452aacf-kube-api-access-427z6\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.437104 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ca1e08-a8cb-42e4-ac54-30f87452aacf-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.589867 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cgjz2"] Oct 04 03:17:10 crc kubenswrapper[4770]: I1004 03:17:10.601290 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cgjz2"] Oct 04 03:17:11 crc kubenswrapper[4770]: I1004 03:17:11.684145 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" path="/var/lib/kubelet/pods/04ca1e08-a8cb-42e4-ac54-30f87452aacf/volumes" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.581863 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw"] Oct 04 03:17:14 crc kubenswrapper[4770]: E1004 03:17:14.582599 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerName="registry-server" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582616 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerName="registry-server" Oct 04 03:17:14 crc kubenswrapper[4770]: E1004 03:17:14.582626 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerName="extract-content" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582633 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerName="extract-content" Oct 04 03:17:14 crc kubenswrapper[4770]: E1004 03:17:14.582648 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerName="extract-utilities" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582655 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerName="extract-utilities" Oct 04 03:17:14 crc kubenswrapper[4770]: E1004 03:17:14.582672 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerName="extract-utilities" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582680 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerName="extract-utilities" Oct 04 03:17:14 crc kubenswrapper[4770]: E1004 03:17:14.582691 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerName="extract-content" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582699 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerName="extract-content" Oct 04 03:17:14 crc kubenswrapper[4770]: E1004 03:17:14.582708 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerName="extract" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582716 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerName="extract" Oct 04 03:17:14 crc kubenswrapper[4770]: E1004 03:17:14.582725 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerName="registry-server" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582732 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerName="registry-server" Oct 04 03:17:14 crc kubenswrapper[4770]: E1004 03:17:14.582746 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerName="util" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582754 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerName="util" Oct 04 03:17:14 crc kubenswrapper[4770]: E1004 03:17:14.582765 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerName="pull" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582772 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerName="pull" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582899 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="04ca1e08-a8cb-42e4-ac54-30f87452aacf" containerName="registry-server" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582919 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="714e499d-6a91-4a5a-b96f-4a56413bc7eb" containerName="extract" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.582929 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbb6aa81-4425-4666-9d9e-7053b90b6981" containerName="registry-server" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.583513 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.587121 4770 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-nvvjs" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.589076 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.589127 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.607170 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw"] Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.693590 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hlcl\" (UniqueName: \"kubernetes.io/projected/2608a80f-0ee6-414c-9564-95b83e22e84e-kube-api-access-9hlcl\") pod \"cert-manager-operator-controller-manager-57cd46d6d-mnhnw\" (UID: \"2608a80f-0ee6-414c-9564-95b83e22e84e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.795066 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hlcl\" (UniqueName: \"kubernetes.io/projected/2608a80f-0ee6-414c-9564-95b83e22e84e-kube-api-access-9hlcl\") pod \"cert-manager-operator-controller-manager-57cd46d6d-mnhnw\" (UID: \"2608a80f-0ee6-414c-9564-95b83e22e84e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.823413 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hlcl\" (UniqueName: \"kubernetes.io/projected/2608a80f-0ee6-414c-9564-95b83e22e84e-kube-api-access-9hlcl\") pod \"cert-manager-operator-controller-manager-57cd46d6d-mnhnw\" (UID: \"2608a80f-0ee6-414c-9564-95b83e22e84e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw" Oct 04 03:17:14 crc kubenswrapper[4770]: I1004 03:17:14.940645 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw" Oct 04 03:17:15 crc kubenswrapper[4770]: I1004 03:17:15.198047 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw"] Oct 04 03:17:15 crc kubenswrapper[4770]: W1004 03:17:15.202362 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2608a80f_0ee6_414c_9564_95b83e22e84e.slice/crio-3026f652a6f656ee7b8d2640e6cfcc6986aae9958d9e3e5f16bd03e8136220b9 WatchSource:0}: Error finding container 3026f652a6f656ee7b8d2640e6cfcc6986aae9958d9e3e5f16bd03e8136220b9: Status 404 returned error can't find the container with id 3026f652a6f656ee7b8d2640e6cfcc6986aae9958d9e3e5f16bd03e8136220b9 Oct 04 03:17:15 crc kubenswrapper[4770]: I1004 03:17:15.293734 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw" event={"ID":"2608a80f-0ee6-414c-9564-95b83e22e84e","Type":"ContainerStarted","Data":"3026f652a6f656ee7b8d2640e6cfcc6986aae9958d9e3e5f16bd03e8136220b9"} Oct 04 03:17:16 crc kubenswrapper[4770]: I1004 03:17:16.665908 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-crfm8" Oct 04 03:17:24 crc kubenswrapper[4770]: I1004 03:17:24.365988 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw" event={"ID":"2608a80f-0ee6-414c-9564-95b83e22e84e","Type":"ContainerStarted","Data":"c4058e3ffd57066a061e6fe7abed56d9a3cace432a7dc95d3e9f1a112c5a91cd"} Oct 04 03:17:24 crc kubenswrapper[4770]: I1004 03:17:24.396088 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-mnhnw" podStartSLOduration=2.441210924 podStartE2EDuration="10.396062893s" podCreationTimestamp="2025-10-04 03:17:14 +0000 UTC" firstStartedPulling="2025-10-04 03:17:15.206196542 +0000 UTC m=+846.498206264" lastFinishedPulling="2025-10-04 03:17:23.161048481 +0000 UTC m=+854.453058233" observedRunningTime="2025-10-04 03:17:24.394481222 +0000 UTC m=+855.686490974" watchObservedRunningTime="2025-10-04 03:17:24.396062893 +0000 UTC m=+855.688072605" Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.755201 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-fwxql"] Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.756694 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.759103 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.759123 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.760681 4770 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-fw7m9" Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.815672 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-fwxql"] Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.878550 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed3e720d-90d9-4f0b-8b40-edf9eda88615-bound-sa-token\") pod \"cert-manager-webhook-d969966f-fwxql\" (UID: \"ed3e720d-90d9-4f0b-8b40-edf9eda88615\") " pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.878804 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs7m9\" (UniqueName: \"kubernetes.io/projected/ed3e720d-90d9-4f0b-8b40-edf9eda88615-kube-api-access-rs7m9\") pod \"cert-manager-webhook-d969966f-fwxql\" (UID: \"ed3e720d-90d9-4f0b-8b40-edf9eda88615\") " pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.980760 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed3e720d-90d9-4f0b-8b40-edf9eda88615-bound-sa-token\") pod \"cert-manager-webhook-d969966f-fwxql\" (UID: \"ed3e720d-90d9-4f0b-8b40-edf9eda88615\") " pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:26 crc kubenswrapper[4770]: I1004 03:17:26.980889 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs7m9\" (UniqueName: \"kubernetes.io/projected/ed3e720d-90d9-4f0b-8b40-edf9eda88615-kube-api-access-rs7m9\") pod \"cert-manager-webhook-d969966f-fwxql\" (UID: \"ed3e720d-90d9-4f0b-8b40-edf9eda88615\") " pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:27 crc kubenswrapper[4770]: I1004 03:17:27.001685 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed3e720d-90d9-4f0b-8b40-edf9eda88615-bound-sa-token\") pod \"cert-manager-webhook-d969966f-fwxql\" (UID: \"ed3e720d-90d9-4f0b-8b40-edf9eda88615\") " pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:27 crc kubenswrapper[4770]: I1004 03:17:27.004543 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs7m9\" (UniqueName: \"kubernetes.io/projected/ed3e720d-90d9-4f0b-8b40-edf9eda88615-kube-api-access-rs7m9\") pod \"cert-manager-webhook-d969966f-fwxql\" (UID: \"ed3e720d-90d9-4f0b-8b40-edf9eda88615\") " pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:27 crc kubenswrapper[4770]: I1004 03:17:27.075750 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:27 crc kubenswrapper[4770]: I1004 03:17:27.370272 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-fwxql"] Oct 04 03:17:27 crc kubenswrapper[4770]: I1004 03:17:27.401122 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-fwxql" event={"ID":"ed3e720d-90d9-4f0b-8b40-edf9eda88615","Type":"ContainerStarted","Data":"88c44b2e2f664c47732c7eca7d3f91f5b13b09f2876aeb7cf856f191074b31f9"} Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.351610 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-298sm"] Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.354396 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.362311 4770 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-sr9vx" Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.367075 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-298sm"] Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.399765 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhdh5\" (UniqueName: \"kubernetes.io/projected/26947e76-92b5-49ad-976f-fb19b0426a3c-kube-api-access-xhdh5\") pod \"cert-manager-cainjector-7d9f95dbf-298sm\" (UID: \"26947e76-92b5-49ad-976f-fb19b0426a3c\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.399904 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/26947e76-92b5-49ad-976f-fb19b0426a3c-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-298sm\" (UID: \"26947e76-92b5-49ad-976f-fb19b0426a3c\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.501558 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhdh5\" (UniqueName: \"kubernetes.io/projected/26947e76-92b5-49ad-976f-fb19b0426a3c-kube-api-access-xhdh5\") pod \"cert-manager-cainjector-7d9f95dbf-298sm\" (UID: \"26947e76-92b5-49ad-976f-fb19b0426a3c\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.501725 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/26947e76-92b5-49ad-976f-fb19b0426a3c-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-298sm\" (UID: \"26947e76-92b5-49ad-976f-fb19b0426a3c\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.532562 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhdh5\" (UniqueName: \"kubernetes.io/projected/26947e76-92b5-49ad-976f-fb19b0426a3c-kube-api-access-xhdh5\") pod \"cert-manager-cainjector-7d9f95dbf-298sm\" (UID: \"26947e76-92b5-49ad-976f-fb19b0426a3c\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.545652 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/26947e76-92b5-49ad-976f-fb19b0426a3c-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-298sm\" (UID: \"26947e76-92b5-49ad-976f-fb19b0426a3c\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.674327 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" Oct 04 03:17:28 crc kubenswrapper[4770]: I1004 03:17:28.985389 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-298sm"] Oct 04 03:17:28 crc kubenswrapper[4770]: W1004 03:17:28.992881 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26947e76_92b5_49ad_976f_fb19b0426a3c.slice/crio-411b03e578fe52c2e7a72dc717335787946de4b6934229c06df087ae6dcf95e5 WatchSource:0}: Error finding container 411b03e578fe52c2e7a72dc717335787946de4b6934229c06df087ae6dcf95e5: Status 404 returned error can't find the container with id 411b03e578fe52c2e7a72dc717335787946de4b6934229c06df087ae6dcf95e5 Oct 04 03:17:29 crc kubenswrapper[4770]: I1004 03:17:29.422527 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" event={"ID":"26947e76-92b5-49ad-976f-fb19b0426a3c","Type":"ContainerStarted","Data":"411b03e578fe52c2e7a72dc717335787946de4b6934229c06df087ae6dcf95e5"} Oct 04 03:17:31 crc kubenswrapper[4770]: I1004 03:17:31.795626 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:17:31 crc kubenswrapper[4770]: I1004 03:17:31.795997 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:17:31 crc kubenswrapper[4770]: I1004 03:17:31.796071 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:17:31 crc kubenswrapper[4770]: I1004 03:17:31.796746 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f8ff4d68e47173d152e933c25298fdeacbf8df65acd72aab55c2b5b7e5c9c52"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:17:31 crc kubenswrapper[4770]: I1004 03:17:31.796801 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://1f8ff4d68e47173d152e933c25298fdeacbf8df65acd72aab55c2b5b7e5c9c52" gracePeriod=600 Oct 04 03:17:32 crc kubenswrapper[4770]: I1004 03:17:32.473612 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="1f8ff4d68e47173d152e933c25298fdeacbf8df65acd72aab55c2b5b7e5c9c52" exitCode=0 Oct 04 03:17:32 crc kubenswrapper[4770]: I1004 03:17:32.474137 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"1f8ff4d68e47173d152e933c25298fdeacbf8df65acd72aab55c2b5b7e5c9c52"} Oct 04 03:17:32 crc kubenswrapper[4770]: I1004 03:17:32.474185 4770 scope.go:117] "RemoveContainer" containerID="d700d3adbc22ca8f48ded121cdd0b6720fb293c613b3f9c60d656bef6938d1c8" Oct 04 03:17:33 crc kubenswrapper[4770]: I1004 03:17:33.482966 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" event={"ID":"26947e76-92b5-49ad-976f-fb19b0426a3c","Type":"ContainerStarted","Data":"93f2c6a4425cf898a8d6ef9a4c311ed649ff003a2f90251f13e8bfb2177d8563"} Oct 04 03:17:33 crc kubenswrapper[4770]: I1004 03:17:33.484624 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-fwxql" event={"ID":"ed3e720d-90d9-4f0b-8b40-edf9eda88615","Type":"ContainerStarted","Data":"258f9b5a01dfccc34540e4a9e2a9a5abbcc25cd0edad280ff223fd50d315c4a7"} Oct 04 03:17:33 crc kubenswrapper[4770]: I1004 03:17:33.484750 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:33 crc kubenswrapper[4770]: I1004 03:17:33.487701 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"73d28cffdd4e71105ae4d9aa35d59b331f61e307981b18b94799778a5b0a5a8d"} Oct 04 03:17:33 crc kubenswrapper[4770]: I1004 03:17:33.503611 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-298sm" podStartSLOduration=2.164610636 podStartE2EDuration="5.503588102s" podCreationTimestamp="2025-10-04 03:17:28 +0000 UTC" firstStartedPulling="2025-10-04 03:17:28.996979305 +0000 UTC m=+860.288989017" lastFinishedPulling="2025-10-04 03:17:32.335956751 +0000 UTC m=+863.627966483" observedRunningTime="2025-10-04 03:17:33.500247034 +0000 UTC m=+864.792256766" watchObservedRunningTime="2025-10-04 03:17:33.503588102 +0000 UTC m=+864.795597814" Oct 04 03:17:33 crc kubenswrapper[4770]: I1004 03:17:33.523109 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-fwxql" podStartSLOduration=2.591496153 podStartE2EDuration="7.523084102s" podCreationTimestamp="2025-10-04 03:17:26 +0000 UTC" firstStartedPulling="2025-10-04 03:17:27.383811925 +0000 UTC m=+858.675821637" lastFinishedPulling="2025-10-04 03:17:32.315399874 +0000 UTC m=+863.607409586" observedRunningTime="2025-10-04 03:17:33.518570283 +0000 UTC m=+864.810580015" watchObservedRunningTime="2025-10-04 03:17:33.523084102 +0000 UTC m=+864.815093824" Oct 04 03:17:37 crc kubenswrapper[4770]: I1004 03:17:37.079356 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-fwxql" Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.504334 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-gwpl4"] Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.505854 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.509228 4770 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-tqr9n" Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.521413 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-gwpl4"] Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.591305 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06c54c3d-d425-46b6-aedb-46606d7e9c0d-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-gwpl4\" (UID: \"06c54c3d-d425-46b6-aedb-46606d7e9c0d\") " pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.591423 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txnb7\" (UniqueName: \"kubernetes.io/projected/06c54c3d-d425-46b6-aedb-46606d7e9c0d-kube-api-access-txnb7\") pod \"cert-manager-7d4cc89fcb-gwpl4\" (UID: \"06c54c3d-d425-46b6-aedb-46606d7e9c0d\") " pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.692299 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txnb7\" (UniqueName: \"kubernetes.io/projected/06c54c3d-d425-46b6-aedb-46606d7e9c0d-kube-api-access-txnb7\") pod \"cert-manager-7d4cc89fcb-gwpl4\" (UID: \"06c54c3d-d425-46b6-aedb-46606d7e9c0d\") " pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.692407 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06c54c3d-d425-46b6-aedb-46606d7e9c0d-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-gwpl4\" (UID: \"06c54c3d-d425-46b6-aedb-46606d7e9c0d\") " pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.720030 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txnb7\" (UniqueName: \"kubernetes.io/projected/06c54c3d-d425-46b6-aedb-46606d7e9c0d-kube-api-access-txnb7\") pod \"cert-manager-7d4cc89fcb-gwpl4\" (UID: \"06c54c3d-d425-46b6-aedb-46606d7e9c0d\") " pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.721464 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06c54c3d-d425-46b6-aedb-46606d7e9c0d-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-gwpl4\" (UID: \"06c54c3d-d425-46b6-aedb-46606d7e9c0d\") " pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" Oct 04 03:17:45 crc kubenswrapper[4770]: I1004 03:17:45.825542 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" Oct 04 03:17:46 crc kubenswrapper[4770]: I1004 03:17:46.125089 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-gwpl4"] Oct 04 03:17:46 crc kubenswrapper[4770]: I1004 03:17:46.596704 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" event={"ID":"06c54c3d-d425-46b6-aedb-46606d7e9c0d","Type":"ContainerStarted","Data":"54c9b1dbeeaa7f3add82686a41da6874d7f08795930a1b65245ccf1365645119"} Oct 04 03:17:46 crc kubenswrapper[4770]: I1004 03:17:46.597353 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" event={"ID":"06c54c3d-d425-46b6-aedb-46606d7e9c0d","Type":"ContainerStarted","Data":"69bb25cd14c76c1853413eefcbb557483e1ba1eefd4242d30217b4370624ef16"} Oct 04 03:17:46 crc kubenswrapper[4770]: I1004 03:17:46.629778 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-gwpl4" podStartSLOduration=1.629747562 podStartE2EDuration="1.629747562s" podCreationTimestamp="2025-10-04 03:17:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:17:46.619931546 +0000 UTC m=+877.911941258" watchObservedRunningTime="2025-10-04 03:17:46.629747562 +0000 UTC m=+877.921757284" Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.475595 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-bkn7s"] Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.477711 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bkn7s" Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.481907 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.483569 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-rqjvp" Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.498000 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.509076 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bkn7s"] Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.585496 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xkbw\" (UniqueName: \"kubernetes.io/projected/d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8-kube-api-access-5xkbw\") pod \"openstack-operator-index-bkn7s\" (UID: \"d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8\") " pod="openstack-operators/openstack-operator-index-bkn7s" Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.687277 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xkbw\" (UniqueName: \"kubernetes.io/projected/d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8-kube-api-access-5xkbw\") pod \"openstack-operator-index-bkn7s\" (UID: \"d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8\") " pod="openstack-operators/openstack-operator-index-bkn7s" Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.719152 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xkbw\" (UniqueName: \"kubernetes.io/projected/d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8-kube-api-access-5xkbw\") pod \"openstack-operator-index-bkn7s\" (UID: \"d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8\") " pod="openstack-operators/openstack-operator-index-bkn7s" Oct 04 03:17:51 crc kubenswrapper[4770]: I1004 03:17:51.803507 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bkn7s" Oct 04 03:17:52 crc kubenswrapper[4770]: I1004 03:17:52.254847 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bkn7s"] Oct 04 03:17:52 crc kubenswrapper[4770]: I1004 03:17:52.647597 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bkn7s" event={"ID":"d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8","Type":"ContainerStarted","Data":"9e93603e7a053477bc1a18644953d62f6d92e39ef065f37fee379c010b3b5abe"} Oct 04 03:17:54 crc kubenswrapper[4770]: I1004 03:17:54.667166 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bkn7s" event={"ID":"d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8","Type":"ContainerStarted","Data":"8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1"} Oct 04 03:17:54 crc kubenswrapper[4770]: I1004 03:17:54.700476 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-bkn7s" podStartSLOduration=2.082222629 podStartE2EDuration="3.700447011s" podCreationTimestamp="2025-10-04 03:17:51 +0000 UTC" firstStartedPulling="2025-10-04 03:17:52.262095074 +0000 UTC m=+883.554104816" lastFinishedPulling="2025-10-04 03:17:53.880319486 +0000 UTC m=+885.172329198" observedRunningTime="2025-10-04 03:17:54.68970191 +0000 UTC m=+885.981711662" watchObservedRunningTime="2025-10-04 03:17:54.700447011 +0000 UTC m=+885.992456773" Oct 04 03:17:54 crc kubenswrapper[4770]: I1004 03:17:54.831596 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-bkn7s"] Oct 04 03:17:55 crc kubenswrapper[4770]: I1004 03:17:55.439340 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qgzdn"] Oct 04 03:17:55 crc kubenswrapper[4770]: I1004 03:17:55.440462 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qgzdn" Oct 04 03:17:55 crc kubenswrapper[4770]: I1004 03:17:55.448431 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qgzdn"] Oct 04 03:17:55 crc kubenswrapper[4770]: I1004 03:17:55.547992 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsx96\" (UniqueName: \"kubernetes.io/projected/9751f79d-b495-4ea5-97ca-a8b4fe50d7ca-kube-api-access-lsx96\") pod \"openstack-operator-index-qgzdn\" (UID: \"9751f79d-b495-4ea5-97ca-a8b4fe50d7ca\") " pod="openstack-operators/openstack-operator-index-qgzdn" Oct 04 03:17:55 crc kubenswrapper[4770]: I1004 03:17:55.649357 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsx96\" (UniqueName: \"kubernetes.io/projected/9751f79d-b495-4ea5-97ca-a8b4fe50d7ca-kube-api-access-lsx96\") pod \"openstack-operator-index-qgzdn\" (UID: \"9751f79d-b495-4ea5-97ca-a8b4fe50d7ca\") " pod="openstack-operators/openstack-operator-index-qgzdn" Oct 04 03:17:55 crc kubenswrapper[4770]: I1004 03:17:55.677097 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsx96\" (UniqueName: \"kubernetes.io/projected/9751f79d-b495-4ea5-97ca-a8b4fe50d7ca-kube-api-access-lsx96\") pod \"openstack-operator-index-qgzdn\" (UID: \"9751f79d-b495-4ea5-97ca-a8b4fe50d7ca\") " pod="openstack-operators/openstack-operator-index-qgzdn" Oct 04 03:17:55 crc kubenswrapper[4770]: I1004 03:17:55.759783 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qgzdn" Oct 04 03:17:56 crc kubenswrapper[4770]: I1004 03:17:56.077925 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qgzdn"] Oct 04 03:17:56 crc kubenswrapper[4770]: I1004 03:17:56.684699 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qgzdn" event={"ID":"9751f79d-b495-4ea5-97ca-a8b4fe50d7ca","Type":"ContainerStarted","Data":"0ddf617c070141d82b640eb8d2d0fa6503dd9bbb4dde25d6986e3d534baa3947"} Oct 04 03:17:56 crc kubenswrapper[4770]: I1004 03:17:56.684906 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-bkn7s" podUID="d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8" containerName="registry-server" containerID="cri-o://8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1" gracePeriod=2 Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.119453 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bkn7s" Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.178585 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xkbw\" (UniqueName: \"kubernetes.io/projected/d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8-kube-api-access-5xkbw\") pod \"d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8\" (UID: \"d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8\") " Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.189124 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8-kube-api-access-5xkbw" (OuterVolumeSpecName: "kube-api-access-5xkbw") pod "d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8" (UID: "d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8"). InnerVolumeSpecName "kube-api-access-5xkbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.281152 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xkbw\" (UniqueName: \"kubernetes.io/projected/d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8-kube-api-access-5xkbw\") on node \"crc\" DevicePath \"\"" Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.696170 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qgzdn" event={"ID":"9751f79d-b495-4ea5-97ca-a8b4fe50d7ca","Type":"ContainerStarted","Data":"6fd6ef621c92f7bb8da9616306ffdc4fd7887fd61c63333e3f74d33f4fc5cfbf"} Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.699815 4770 generic.go:334] "Generic (PLEG): container finished" podID="d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8" containerID="8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1" exitCode=0 Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.699904 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bkn7s" event={"ID":"d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8","Type":"ContainerDied","Data":"8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1"} Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.699923 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bkn7s" Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.699954 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bkn7s" event={"ID":"d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8","Type":"ContainerDied","Data":"9e93603e7a053477bc1a18644953d62f6d92e39ef065f37fee379c010b3b5abe"} Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.699992 4770 scope.go:117] "RemoveContainer" containerID="8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1" Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.724568 4770 scope.go:117] "RemoveContainer" containerID="8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1" Oct 04 03:17:57 crc kubenswrapper[4770]: E1004 03:17:57.725530 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1\": container with ID starting with 8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1 not found: ID does not exist" containerID="8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1" Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.725628 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1"} err="failed to get container status \"8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1\": rpc error: code = NotFound desc = could not find container \"8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1\": container with ID starting with 8cfbdd58a97fa0e8db0b0543a7e2fff7717405c3412853f71c5513e7cb6dc5f1 not found: ID does not exist" Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.730365 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qgzdn" podStartSLOduration=2.158622826 podStartE2EDuration="2.730335455s" podCreationTimestamp="2025-10-04 03:17:55 +0000 UTC" firstStartedPulling="2025-10-04 03:17:56.093704251 +0000 UTC m=+887.385713983" lastFinishedPulling="2025-10-04 03:17:56.66541686 +0000 UTC m=+887.957426612" observedRunningTime="2025-10-04 03:17:57.724956984 +0000 UTC m=+889.016966726" watchObservedRunningTime="2025-10-04 03:17:57.730335455 +0000 UTC m=+889.022345167" Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.744038 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-bkn7s"] Oct 04 03:17:57 crc kubenswrapper[4770]: I1004 03:17:57.754619 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-bkn7s"] Oct 04 03:17:59 crc kubenswrapper[4770]: I1004 03:17:59.683504 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8" path="/var/lib/kubelet/pods/d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8/volumes" Oct 04 03:18:05 crc kubenswrapper[4770]: I1004 03:18:05.760606 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-qgzdn" Oct 04 03:18:05 crc kubenswrapper[4770]: I1004 03:18:05.761893 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-qgzdn" Oct 04 03:18:05 crc kubenswrapper[4770]: I1004 03:18:05.816330 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-qgzdn" Oct 04 03:18:05 crc kubenswrapper[4770]: I1004 03:18:05.869777 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-qgzdn" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.494578 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj"] Oct 04 03:18:12 crc kubenswrapper[4770]: E1004 03:18:12.495348 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8" containerName="registry-server" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.495363 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8" containerName="registry-server" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.495482 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e06ad1-eeda-4ee9-b3f7-40061ed6b9a8" containerName="registry-server" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.496367 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.498958 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-24xtg" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.522750 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj"] Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.639116 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-util\") pod \"41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.639469 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-bundle\") pod \"41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.639580 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p79rb\" (UniqueName: \"kubernetes.io/projected/7d1c550a-15df-4478-adf8-eb2a329fa625-kube-api-access-p79rb\") pod \"41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.740823 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-bundle\") pod \"41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.740968 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p79rb\" (UniqueName: \"kubernetes.io/projected/7d1c550a-15df-4478-adf8-eb2a329fa625-kube-api-access-p79rb\") pod \"41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.741248 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-util\") pod \"41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.742763 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-bundle\") pod \"41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.742998 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-util\") pod \"41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.773294 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p79rb\" (UniqueName: \"kubernetes.io/projected/7d1c550a-15df-4478-adf8-eb2a329fa625-kube-api-access-p79rb\") pod \"41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:12 crc kubenswrapper[4770]: I1004 03:18:12.814138 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:13 crc kubenswrapper[4770]: I1004 03:18:13.330647 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj"] Oct 04 03:18:13 crc kubenswrapper[4770]: W1004 03:18:13.345049 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d1c550a_15df_4478_adf8_eb2a329fa625.slice/crio-fdb2e2a318f5a82c316d9300a1cbb5713e8cc28fb8d1f0a08da82bea3fea8950 WatchSource:0}: Error finding container fdb2e2a318f5a82c316d9300a1cbb5713e8cc28fb8d1f0a08da82bea3fea8950: Status 404 returned error can't find the container with id fdb2e2a318f5a82c316d9300a1cbb5713e8cc28fb8d1f0a08da82bea3fea8950 Oct 04 03:18:13 crc kubenswrapper[4770]: I1004 03:18:13.848659 4770 generic.go:334] "Generic (PLEG): container finished" podID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerID="2788a1cd8cf26935b347b26c9dd9fe4fa7099e007f7ef439253114906a3f9c72" exitCode=0 Oct 04 03:18:13 crc kubenswrapper[4770]: I1004 03:18:13.848727 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" event={"ID":"7d1c550a-15df-4478-adf8-eb2a329fa625","Type":"ContainerDied","Data":"2788a1cd8cf26935b347b26c9dd9fe4fa7099e007f7ef439253114906a3f9c72"} Oct 04 03:18:13 crc kubenswrapper[4770]: I1004 03:18:13.848766 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" event={"ID":"7d1c550a-15df-4478-adf8-eb2a329fa625","Type":"ContainerStarted","Data":"fdb2e2a318f5a82c316d9300a1cbb5713e8cc28fb8d1f0a08da82bea3fea8950"} Oct 04 03:18:15 crc kubenswrapper[4770]: I1004 03:18:15.871902 4770 generic.go:334] "Generic (PLEG): container finished" podID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerID="9ebd395940fdf56b898e832651df6368f02c4f3951540b9f806c8061f318b79e" exitCode=0 Oct 04 03:18:15 crc kubenswrapper[4770]: I1004 03:18:15.872053 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" event={"ID":"7d1c550a-15df-4478-adf8-eb2a329fa625","Type":"ContainerDied","Data":"9ebd395940fdf56b898e832651df6368f02c4f3951540b9f806c8061f318b79e"} Oct 04 03:18:16 crc kubenswrapper[4770]: I1004 03:18:16.884238 4770 generic.go:334] "Generic (PLEG): container finished" podID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerID="966a39316b26fe7359c65603891d8aba6dad0b19facb700b45e415632ac9a5f2" exitCode=0 Oct 04 03:18:16 crc kubenswrapper[4770]: I1004 03:18:16.884309 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" event={"ID":"7d1c550a-15df-4478-adf8-eb2a329fa625","Type":"ContainerDied","Data":"966a39316b26fe7359c65603891d8aba6dad0b19facb700b45e415632ac9a5f2"} Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.243277 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.327081 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-bundle\") pod \"7d1c550a-15df-4478-adf8-eb2a329fa625\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.327281 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-util\") pod \"7d1c550a-15df-4478-adf8-eb2a329fa625\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.327423 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p79rb\" (UniqueName: \"kubernetes.io/projected/7d1c550a-15df-4478-adf8-eb2a329fa625-kube-api-access-p79rb\") pod \"7d1c550a-15df-4478-adf8-eb2a329fa625\" (UID: \"7d1c550a-15df-4478-adf8-eb2a329fa625\") " Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.328518 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-bundle" (OuterVolumeSpecName: "bundle") pod "7d1c550a-15df-4478-adf8-eb2a329fa625" (UID: "7d1c550a-15df-4478-adf8-eb2a329fa625"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.334272 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d1c550a-15df-4478-adf8-eb2a329fa625-kube-api-access-p79rb" (OuterVolumeSpecName: "kube-api-access-p79rb") pod "7d1c550a-15df-4478-adf8-eb2a329fa625" (UID: "7d1c550a-15df-4478-adf8-eb2a329fa625"). InnerVolumeSpecName "kube-api-access-p79rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.353581 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-util" (OuterVolumeSpecName: "util") pod "7d1c550a-15df-4478-adf8-eb2a329fa625" (UID: "7d1c550a-15df-4478-adf8-eb2a329fa625"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.428711 4770 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.428784 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p79rb\" (UniqueName: \"kubernetes.io/projected/7d1c550a-15df-4478-adf8-eb2a329fa625-kube-api-access-p79rb\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.428804 4770 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d1c550a-15df-4478-adf8-eb2a329fa625-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.905922 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" event={"ID":"7d1c550a-15df-4478-adf8-eb2a329fa625","Type":"ContainerDied","Data":"fdb2e2a318f5a82c316d9300a1cbb5713e8cc28fb8d1f0a08da82bea3fea8950"} Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.905976 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdb2e2a318f5a82c316d9300a1cbb5713e8cc28fb8d1f0a08da82bea3fea8950" Oct 04 03:18:18 crc kubenswrapper[4770]: I1004 03:18:18.906197 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.669709 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll"] Oct 04 03:18:25 crc kubenswrapper[4770]: E1004 03:18:25.670748 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerName="extract" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.670765 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerName="extract" Oct 04 03:18:25 crc kubenswrapper[4770]: E1004 03:18:25.670800 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerName="pull" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.670809 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerName="pull" Oct 04 03:18:25 crc kubenswrapper[4770]: E1004 03:18:25.670821 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerName="util" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.670829 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerName="util" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.670961 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d1c550a-15df-4478-adf8-eb2a329fa625" containerName="extract" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.671706 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.682405 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-5znb9" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.717281 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll"] Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.743731 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2pfl\" (UniqueName: \"kubernetes.io/projected/c2be3a13-80bd-4d80-9f19-35483ae16fc4-kube-api-access-c2pfl\") pod \"openstack-operator-controller-operator-cd5c57ff8-fdfll\" (UID: \"c2be3a13-80bd-4d80-9f19-35483ae16fc4\") " pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.845821 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2pfl\" (UniqueName: \"kubernetes.io/projected/c2be3a13-80bd-4d80-9f19-35483ae16fc4-kube-api-access-c2pfl\") pod \"openstack-operator-controller-operator-cd5c57ff8-fdfll\" (UID: \"c2be3a13-80bd-4d80-9f19-35483ae16fc4\") " pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.932225 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2pfl\" (UniqueName: \"kubernetes.io/projected/c2be3a13-80bd-4d80-9f19-35483ae16fc4-kube-api-access-c2pfl\") pod \"openstack-operator-controller-operator-cd5c57ff8-fdfll\" (UID: \"c2be3a13-80bd-4d80-9f19-35483ae16fc4\") " pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" Oct 04 03:18:25 crc kubenswrapper[4770]: I1004 03:18:25.995044 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" Oct 04 03:18:26 crc kubenswrapper[4770]: I1004 03:18:26.249520 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll"] Oct 04 03:18:26 crc kubenswrapper[4770]: I1004 03:18:26.975107 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" event={"ID":"c2be3a13-80bd-4d80-9f19-35483ae16fc4","Type":"ContainerStarted","Data":"413332292b4229d2ed63b3a1dd4c427892603740065fbdfa5ae09a46dd19d5a7"} Oct 04 03:18:31 crc kubenswrapper[4770]: I1004 03:18:31.007534 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" event={"ID":"c2be3a13-80bd-4d80-9f19-35483ae16fc4","Type":"ContainerStarted","Data":"cba12183bd62c77369fbf377c08576e549750e5feefeb184b4cb950c9620d2c3"} Oct 04 03:18:34 crc kubenswrapper[4770]: I1004 03:18:34.039253 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" event={"ID":"c2be3a13-80bd-4d80-9f19-35483ae16fc4","Type":"ContainerStarted","Data":"140ab05b0c18c4116e3bc1cbf075c10826c476f00da3a3d411b0dea28a40e81a"} Oct 04 03:18:34 crc kubenswrapper[4770]: I1004 03:18:34.039919 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" Oct 04 03:18:34 crc kubenswrapper[4770]: I1004 03:18:34.099034 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" podStartSLOduration=1.7717211050000001 podStartE2EDuration="9.099018395s" podCreationTimestamp="2025-10-04 03:18:25 +0000 UTC" firstStartedPulling="2025-10-04 03:18:26.260763135 +0000 UTC m=+917.552772867" lastFinishedPulling="2025-10-04 03:18:33.588060445 +0000 UTC m=+924.880070157" observedRunningTime="2025-10-04 03:18:34.095673357 +0000 UTC m=+925.387683069" watchObservedRunningTime="2025-10-04 03:18:34.099018395 +0000 UTC m=+925.391028107" Oct 04 03:18:35 crc kubenswrapper[4770]: I1004 03:18:35.051077 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-cd5c57ff8-fdfll" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.250993 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.252866 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.259971 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.260844 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.271510 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.272916 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.274078 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.281983 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-lrpgk" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.282461 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-fdsq9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.282592 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-2bgbb" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.283529 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.299939 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.312985 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.314314 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.317710 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-bvczd" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.329807 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnsgp\" (UniqueName: \"kubernetes.io/projected/7a0cc679-8960-4bae-9da9-f790f949b5bf-kube-api-access-jnsgp\") pod \"designate-operator-controller-manager-58d86cd59d-qj4zt\" (UID: \"7a0cc679-8960-4bae-9da9-f790f949b5bf\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.329844 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs2jb\" (UniqueName: \"kubernetes.io/projected/909aa645-88e5-4e67-835e-34854c2f5a74-kube-api-access-xs2jb\") pod \"barbican-operator-controller-manager-5b974f6766-nq6dg\" (UID: \"909aa645-88e5-4e67-835e-34854c2f5a74\") " pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.329914 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkwsm\" (UniqueName: \"kubernetes.io/projected/c0e405f3-b246-4741-8580-248db52f995d-kube-api-access-xkwsm\") pod \"cinder-operator-controller-manager-84bd8f6848-m269q\" (UID: \"c0e405f3-b246-4741-8580-248db52f995d\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.333664 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.336568 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.338243 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.343776 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-24wdm" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.354453 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.360536 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6675647785-zkw64"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.416523 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.427094 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-j4mph" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.428388 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6675647785-zkw64"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.431100 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnlvz\" (UniqueName: \"kubernetes.io/projected/aab021e2-b955-4efa-8668-1444f506ca5d-kube-api-access-lnlvz\") pod \"heat-operator-controller-manager-5c497dbdb-bcfrt\" (UID: \"aab021e2-b955-4efa-8668-1444f506ca5d\") " pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.431170 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkwsm\" (UniqueName: \"kubernetes.io/projected/c0e405f3-b246-4741-8580-248db52f995d-kube-api-access-xkwsm\") pod \"cinder-operator-controller-manager-84bd8f6848-m269q\" (UID: \"c0e405f3-b246-4741-8580-248db52f995d\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.431229 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnsgp\" (UniqueName: \"kubernetes.io/projected/7a0cc679-8960-4bae-9da9-f790f949b5bf-kube-api-access-jnsgp\") pod \"designate-operator-controller-manager-58d86cd59d-qj4zt\" (UID: \"7a0cc679-8960-4bae-9da9-f790f949b5bf\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.431248 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs2jb\" (UniqueName: \"kubernetes.io/projected/909aa645-88e5-4e67-835e-34854c2f5a74-kube-api-access-xs2jb\") pod \"barbican-operator-controller-manager-5b974f6766-nq6dg\" (UID: \"909aa645-88e5-4e67-835e-34854c2f5a74\") " pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.431273 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wthqn\" (UniqueName: \"kubernetes.io/projected/e26cfeef-31d4-452e-bbf6-f54ab0d3b331-kube-api-access-wthqn\") pod \"glance-operator-controller-manager-698456cdc6-dzkch\" (UID: \"e26cfeef-31d4-452e-bbf6-f54ab0d3b331\") " pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.446257 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.447891 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.450318 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.451481 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-n8287" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.465141 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.472804 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkwsm\" (UniqueName: \"kubernetes.io/projected/c0e405f3-b246-4741-8580-248db52f995d-kube-api-access-xkwsm\") pod \"cinder-operator-controller-manager-84bd8f6848-m269q\" (UID: \"c0e405f3-b246-4741-8580-248db52f995d\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.478644 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnsgp\" (UniqueName: \"kubernetes.io/projected/7a0cc679-8960-4bae-9da9-f790f949b5bf-kube-api-access-jnsgp\") pod \"designate-operator-controller-manager-58d86cd59d-qj4zt\" (UID: \"7a0cc679-8960-4bae-9da9-f790f949b5bf\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.489080 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.490318 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.494170 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-phx7g" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.511963 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs2jb\" (UniqueName: \"kubernetes.io/projected/909aa645-88e5-4e67-835e-34854c2f5a74-kube-api-access-xs2jb\") pod \"barbican-operator-controller-manager-5b974f6766-nq6dg\" (UID: \"909aa645-88e5-4e67-835e-34854c2f5a74\") " pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.538344 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wthqn\" (UniqueName: \"kubernetes.io/projected/e26cfeef-31d4-452e-bbf6-f54ab0d3b331-kube-api-access-wthqn\") pod \"glance-operator-controller-manager-698456cdc6-dzkch\" (UID: \"e26cfeef-31d4-452e-bbf6-f54ab0d3b331\") " pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.538402 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7cgp\" (UniqueName: \"kubernetes.io/projected/1fbc59ec-b29e-4843-b34c-43335a590c11-kube-api-access-q7cgp\") pod \"infra-operator-controller-manager-84788b6bc5-9tpl9\" (UID: \"1fbc59ec-b29e-4843-b34c-43335a590c11\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.538471 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnlvz\" (UniqueName: \"kubernetes.io/projected/aab021e2-b955-4efa-8668-1444f506ca5d-kube-api-access-lnlvz\") pod \"heat-operator-controller-manager-5c497dbdb-bcfrt\" (UID: \"aab021e2-b955-4efa-8668-1444f506ca5d\") " pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.538645 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td7sq\" (UniqueName: \"kubernetes.io/projected/72f16f95-3791-461e-8d91-d89005030236-kube-api-access-td7sq\") pod \"ironic-operator-controller-manager-6f5894c49f-56p4n\" (UID: \"72f16f95-3791-461e-8d91-d89005030236\") " pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.538715 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dng89\" (UniqueName: \"kubernetes.io/projected/d35d4909-8960-4d33-9235-9692bf3a4edd-kube-api-access-dng89\") pod \"horizon-operator-controller-manager-6675647785-zkw64\" (UID: \"d35d4909-8960-4d33-9235-9692bf3a4edd\") " pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.538760 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1fbc59ec-b29e-4843-b34c-43335a590c11-cert\") pod \"infra-operator-controller-manager-84788b6bc5-9tpl9\" (UID: \"1fbc59ec-b29e-4843-b34c-43335a590c11\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.541598 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.543190 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.550286 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-wl6b9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.562956 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.574350 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.587832 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.588758 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.589717 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.591030 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnlvz\" (UniqueName: \"kubernetes.io/projected/aab021e2-b955-4efa-8668-1444f506ca5d-kube-api-access-lnlvz\") pod \"heat-operator-controller-manager-5c497dbdb-bcfrt\" (UID: \"aab021e2-b955-4efa-8668-1444f506ca5d\") " pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.608883 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-vmm7r" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.608934 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.609050 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wthqn\" (UniqueName: \"kubernetes.io/projected/e26cfeef-31d4-452e-bbf6-f54ab0d3b331-kube-api-access-wthqn\") pod \"glance-operator-controller-manager-698456cdc6-dzkch\" (UID: \"e26cfeef-31d4-452e-bbf6-f54ab0d3b331\") " pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.618239 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.621581 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.634601 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.641047 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7cgp\" (UniqueName: \"kubernetes.io/projected/1fbc59ec-b29e-4843-b34c-43335a590c11-kube-api-access-q7cgp\") pod \"infra-operator-controller-manager-84788b6bc5-9tpl9\" (UID: \"1fbc59ec-b29e-4843-b34c-43335a590c11\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.641133 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q944g\" (UniqueName: \"kubernetes.io/projected/279505bc-0485-4aa9-87ce-930272978c29-kube-api-access-q944g\") pod \"manila-operator-controller-manager-7cb48dbc-rwskr\" (UID: \"279505bc-0485-4aa9-87ce-930272978c29\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.641186 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td7sq\" (UniqueName: \"kubernetes.io/projected/72f16f95-3791-461e-8d91-d89005030236-kube-api-access-td7sq\") pod \"ironic-operator-controller-manager-6f5894c49f-56p4n\" (UID: \"72f16f95-3791-461e-8d91-d89005030236\") " pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.641221 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dng89\" (UniqueName: \"kubernetes.io/projected/d35d4909-8960-4d33-9235-9692bf3a4edd-kube-api-access-dng89\") pod \"horizon-operator-controller-manager-6675647785-zkw64\" (UID: \"d35d4909-8960-4d33-9235-9692bf3a4edd\") " pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.641250 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d294s\" (UniqueName: \"kubernetes.io/projected/c1bba079-ef51-40bb-b92d-b03a6e31d457-kube-api-access-d294s\") pod \"keystone-operator-controller-manager-57c9cdcf57-95452\" (UID: \"c1bba079-ef51-40bb-b92d-b03a6e31d457\") " pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.641271 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1fbc59ec-b29e-4843-b34c-43335a590c11-cert\") pod \"infra-operator-controller-manager-84788b6bc5-9tpl9\" (UID: \"1fbc59ec-b29e-4843-b34c-43335a590c11\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.648653 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.650366 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.653312 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qs9wh" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.666570 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1fbc59ec-b29e-4843-b34c-43335a590c11-cert\") pod \"infra-operator-controller-manager-84788b6bc5-9tpl9\" (UID: \"1fbc59ec-b29e-4843-b34c-43335a590c11\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.666652 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.666815 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7cgp\" (UniqueName: \"kubernetes.io/projected/1fbc59ec-b29e-4843-b34c-43335a590c11-kube-api-access-q7cgp\") pod \"infra-operator-controller-manager-84788b6bc5-9tpl9\" (UID: \"1fbc59ec-b29e-4843-b34c-43335a590c11\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.673577 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.673849 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dng89\" (UniqueName: \"kubernetes.io/projected/d35d4909-8960-4d33-9235-9692bf3a4edd-kube-api-access-dng89\") pod \"horizon-operator-controller-manager-6675647785-zkw64\" (UID: \"d35d4909-8960-4d33-9235-9692bf3a4edd\") " pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.676620 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.679345 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td7sq\" (UniqueName: \"kubernetes.io/projected/72f16f95-3791-461e-8d91-d89005030236-kube-api-access-td7sq\") pod \"ironic-operator-controller-manager-6f5894c49f-56p4n\" (UID: \"72f16f95-3791-461e-8d91-d89005030236\") " pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.679421 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-zv67z" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.705095 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.716918 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.721060 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.724538 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-c27lb" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.729834 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.731495 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.733081 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-lttpz" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.742233 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q944g\" (UniqueName: \"kubernetes.io/projected/279505bc-0485-4aa9-87ce-930272978c29-kube-api-access-q944g\") pod \"manila-operator-controller-manager-7cb48dbc-rwskr\" (UID: \"279505bc-0485-4aa9-87ce-930272978c29\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.742324 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhm7r\" (UniqueName: \"kubernetes.io/projected/acfbbcdb-767c-4a41-924e-dc60bbd3a5ae-kube-api-access-rhm7r\") pod \"neutron-operator-controller-manager-69b956fbf6-rn8fr\" (UID: \"acfbbcdb-767c-4a41-924e-dc60bbd3a5ae\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.742348 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d294s\" (UniqueName: \"kubernetes.io/projected/c1bba079-ef51-40bb-b92d-b03a6e31d457-kube-api-access-d294s\") pod \"keystone-operator-controller-manager-57c9cdcf57-95452\" (UID: \"c1bba079-ef51-40bb-b92d-b03a6e31d457\") " pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.742388 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blk4q\" (UniqueName: \"kubernetes.io/projected/09622ed2-b879-4add-9df7-0fb77dfa85cb-kube-api-access-blk4q\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-67q4w\" (UID: \"09622ed2-b879-4add-9df7-0fb77dfa85cb\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.754212 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.764268 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q944g\" (UniqueName: \"kubernetes.io/projected/279505bc-0485-4aa9-87ce-930272978c29-kube-api-access-q944g\") pod \"manila-operator-controller-manager-7cb48dbc-rwskr\" (UID: \"279505bc-0485-4aa9-87ce-930272978c29\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.767281 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.768835 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.773226 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d294s\" (UniqueName: \"kubernetes.io/projected/c1bba079-ef51-40bb-b92d-b03a6e31d457-kube-api-access-d294s\") pod \"keystone-operator-controller-manager-57c9cdcf57-95452\" (UID: \"c1bba079-ef51-40bb-b92d-b03a6e31d457\") " pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.779555 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.785323 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.786544 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.789722 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-bz4br" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.790117 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.790275 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.791235 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.792993 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-jxmmh" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.800098 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.801888 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.803086 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.803827 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-6gt8m" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.814602 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.823769 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.824669 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.851663 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd7h6\" (UniqueName: \"kubernetes.io/projected/6231ce27-f748-4e48-ac46-60b1d336e153-kube-api-access-jd7h6\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6\" (UID: \"6231ce27-f748-4e48-ac46-60b1d336e153\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.851718 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtngz\" (UniqueName: \"kubernetes.io/projected/fcfac3b9-2e3b-453e-b434-42159c1f92a9-kube-api-access-qtngz\") pod \"nova-operator-controller-manager-6c9b57c67-sn2wq\" (UID: \"fcfac3b9-2e3b-453e-b434-42159c1f92a9\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.851784 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhm7r\" (UniqueName: \"kubernetes.io/projected/acfbbcdb-767c-4a41-924e-dc60bbd3a5ae-kube-api-access-rhm7r\") pod \"neutron-operator-controller-manager-69b956fbf6-rn8fr\" (UID: \"acfbbcdb-767c-4a41-924e-dc60bbd3a5ae\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.851806 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsgxw\" (UniqueName: \"kubernetes.io/projected/c0126071-94d2-4c21-ac54-ea6de14946ef-kube-api-access-hsgxw\") pod \"octavia-operator-controller-manager-69f59f9d8-fkh9b\" (UID: \"c0126071-94d2-4c21-ac54-ea6de14946ef\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.851824 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6231ce27-f748-4e48-ac46-60b1d336e153-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6\" (UID: \"6231ce27-f748-4e48-ac46-60b1d336e153\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.851845 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wk76\" (UniqueName: \"kubernetes.io/projected/718d82f1-fa64-4a9d-8780-318b056f44a0-kube-api-access-4wk76\") pod \"placement-operator-controller-manager-66f6d6849b-x8ljn\" (UID: \"718d82f1-fa64-4a9d-8780-318b056f44a0\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.851870 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blk4q\" (UniqueName: \"kubernetes.io/projected/09622ed2-b879-4add-9df7-0fb77dfa85cb-kube-api-access-blk4q\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-67q4w\" (UID: \"09622ed2-b879-4add-9df7-0fb77dfa85cb\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.851898 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-852t7\" (UniqueName: \"kubernetes.io/projected/2b8720ec-1142-490f-9a6e-9c94e08db137-kube-api-access-852t7\") pod \"ovn-operator-controller-manager-c968bb45-bd27z\" (UID: \"2b8720ec-1142-490f-9a6e-9c94e08db137\") " pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.852638 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.861780 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.862934 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.866476 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.867296 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-fbwnk" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.877435 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.889269 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blk4q\" (UniqueName: \"kubernetes.io/projected/09622ed2-b879-4add-9df7-0fb77dfa85cb-kube-api-access-blk4q\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-67q4w\" (UID: \"09622ed2-b879-4add-9df7-0fb77dfa85cb\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.904455 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.915790 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-rsjnd" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.931847 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz"] Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.938900 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhm7r\" (UniqueName: \"kubernetes.io/projected/acfbbcdb-767c-4a41-924e-dc60bbd3a5ae-kube-api-access-rhm7r\") pod \"neutron-operator-controller-manager-69b956fbf6-rn8fr\" (UID: \"acfbbcdb-767c-4a41-924e-dc60bbd3a5ae\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.953237 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-852t7\" (UniqueName: \"kubernetes.io/projected/2b8720ec-1142-490f-9a6e-9c94e08db137-kube-api-access-852t7\") pod \"ovn-operator-controller-manager-c968bb45-bd27z\" (UID: \"2b8720ec-1142-490f-9a6e-9c94e08db137\") " pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.953292 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk2cr\" (UniqueName: \"kubernetes.io/projected/34a19c5f-604f-4225-be1d-c66d51b0919d-kube-api-access-gk2cr\") pod \"swift-operator-controller-manager-76d5577b-8d8z6\" (UID: \"34a19c5f-604f-4225-be1d-c66d51b0919d\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.953324 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd7h6\" (UniqueName: \"kubernetes.io/projected/6231ce27-f748-4e48-ac46-60b1d336e153-kube-api-access-jd7h6\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6\" (UID: \"6231ce27-f748-4e48-ac46-60b1d336e153\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.953361 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtngz\" (UniqueName: \"kubernetes.io/projected/fcfac3b9-2e3b-453e-b434-42159c1f92a9-kube-api-access-qtngz\") pod \"nova-operator-controller-manager-6c9b57c67-sn2wq\" (UID: \"fcfac3b9-2e3b-453e-b434-42159c1f92a9\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.953385 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kccrp\" (UniqueName: \"kubernetes.io/projected/48b4bf81-4d46-4b8c-8a98-2112cb67104c-kube-api-access-kccrp\") pod \"telemetry-operator-controller-manager-f589c7597-26xxz\" (UID: \"48b4bf81-4d46-4b8c-8a98-2112cb67104c\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.953436 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsgxw\" (UniqueName: \"kubernetes.io/projected/c0126071-94d2-4c21-ac54-ea6de14946ef-kube-api-access-hsgxw\") pod \"octavia-operator-controller-manager-69f59f9d8-fkh9b\" (UID: \"c0126071-94d2-4c21-ac54-ea6de14946ef\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.953456 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6231ce27-f748-4e48-ac46-60b1d336e153-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6\" (UID: \"6231ce27-f748-4e48-ac46-60b1d336e153\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.953475 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wk76\" (UniqueName: \"kubernetes.io/projected/718d82f1-fa64-4a9d-8780-318b056f44a0-kube-api-access-4wk76\") pod \"placement-operator-controller-manager-66f6d6849b-x8ljn\" (UID: \"718d82f1-fa64-4a9d-8780-318b056f44a0\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" Oct 04 03:19:12 crc kubenswrapper[4770]: E1004 03:19:12.954427 4770 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 03:19:12 crc kubenswrapper[4770]: E1004 03:19:12.954545 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6231ce27-f748-4e48-ac46-60b1d336e153-cert podName:6231ce27-f748-4e48-ac46-60b1d336e153 nodeName:}" failed. No retries permitted until 2025-10-04 03:19:13.45445978 +0000 UTC m=+964.746469492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6231ce27-f748-4e48-ac46-60b1d336e153-cert") pod "openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" (UID: "6231ce27-f748-4e48-ac46-60b1d336e153") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 03:19:12 crc kubenswrapper[4770]: I1004 03:19:12.968855 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6"] Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.010516 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsgxw\" (UniqueName: \"kubernetes.io/projected/c0126071-94d2-4c21-ac54-ea6de14946ef-kube-api-access-hsgxw\") pod \"octavia-operator-controller-manager-69f59f9d8-fkh9b\" (UID: \"c0126071-94d2-4c21-ac54-ea6de14946ef\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.016902 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.026178 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-852t7\" (UniqueName: \"kubernetes.io/projected/2b8720ec-1142-490f-9a6e-9c94e08db137-kube-api-access-852t7\") pod \"ovn-operator-controller-manager-c968bb45-bd27z\" (UID: \"2b8720ec-1142-490f-9a6e-9c94e08db137\") " pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.031725 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.034859 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd7h6\" (UniqueName: \"kubernetes.io/projected/6231ce27-f748-4e48-ac46-60b1d336e153-kube-api-access-jd7h6\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6\" (UID: \"6231ce27-f748-4e48-ac46-60b1d336e153\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.037852 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtngz\" (UniqueName: \"kubernetes.io/projected/fcfac3b9-2e3b-453e-b434-42159c1f92a9-kube-api-access-qtngz\") pod \"nova-operator-controller-manager-6c9b57c67-sn2wq\" (UID: \"fcfac3b9-2e3b-453e-b434-42159c1f92a9\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.041108 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4"] Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.042927 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.048652 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.048750 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-ck5wq" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.055898 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4"] Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.056933 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kccrp\" (UniqueName: \"kubernetes.io/projected/48b4bf81-4d46-4b8c-8a98-2112cb67104c-kube-api-access-kccrp\") pod \"telemetry-operator-controller-manager-f589c7597-26xxz\" (UID: \"48b4bf81-4d46-4b8c-8a98-2112cb67104c\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.057071 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk2cr\" (UniqueName: \"kubernetes.io/projected/34a19c5f-604f-4225-be1d-c66d51b0919d-kube-api-access-gk2cr\") pod \"swift-operator-controller-manager-76d5577b-8d8z6\" (UID: \"34a19c5f-604f-4225-be1d-c66d51b0919d\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.061380 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.091077 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.129947 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p"] Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.142581 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kccrp\" (UniqueName: \"kubernetes.io/projected/48b4bf81-4d46-4b8c-8a98-2112cb67104c-kube-api-access-kccrp\") pod \"telemetry-operator-controller-manager-f589c7597-26xxz\" (UID: \"48b4bf81-4d46-4b8c-8a98-2112cb67104c\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.143843 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.147682 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wk76\" (UniqueName: \"kubernetes.io/projected/718d82f1-fa64-4a9d-8780-318b056f44a0-kube-api-access-4wk76\") pod \"placement-operator-controller-manager-66f6d6849b-x8ljn\" (UID: \"718d82f1-fa64-4a9d-8780-318b056f44a0\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.149663 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-p27jg" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.153829 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk2cr\" (UniqueName: \"kubernetes.io/projected/34a19c5f-604f-4225-be1d-c66d51b0919d-kube-api-access-gk2cr\") pod \"swift-operator-controller-manager-76d5577b-8d8z6\" (UID: \"34a19c5f-604f-4225-be1d-c66d51b0919d\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.162185 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kv9s\" (UniqueName: \"kubernetes.io/projected/64c06e35-b80f-4889-a680-36c627c99a06-kube-api-access-7kv9s\") pod \"test-operator-controller-manager-6bb6dcddc-mhcm4\" (UID: \"64c06e35-b80f-4889-a680-36c627c99a06\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.162380 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p"] Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.223812 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4"] Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.225573 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.241636 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.241936 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-blf8j" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.243450 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4"] Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.271528 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvd9b\" (UniqueName: \"kubernetes.io/projected/059b3d3c-de5a-490e-94ff-f15b249de3b3-kube-api-access-fvd9b\") pod \"watcher-operator-controller-manager-5d98cc5575-l5k6p\" (UID: \"059b3d3c-de5a-490e-94ff-f15b249de3b3\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.271608 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c-cert\") pod \"openstack-operator-controller-manager-7996577b9d-7kzr4\" (UID: \"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c\") " pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.271656 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kv9s\" (UniqueName: \"kubernetes.io/projected/64c06e35-b80f-4889-a680-36c627c99a06-kube-api-access-7kv9s\") pod \"test-operator-controller-manager-6bb6dcddc-mhcm4\" (UID: \"64c06e35-b80f-4889-a680-36c627c99a06\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.271693 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7w8s\" (UniqueName: \"kubernetes.io/projected/ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c-kube-api-access-s7w8s\") pod \"openstack-operator-controller-manager-7996577b9d-7kzr4\" (UID: \"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c\") " pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.284022 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.295195 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.300731 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv"] Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.302514 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.306778 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-dvp27" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.309788 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kv9s\" (UniqueName: \"kubernetes.io/projected/64c06e35-b80f-4889-a680-36c627c99a06-kube-api-access-7kv9s\") pod \"test-operator-controller-manager-6bb6dcddc-mhcm4\" (UID: \"64c06e35-b80f-4889-a680-36c627c99a06\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.319981 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv"] Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.328857 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.342897 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.368512 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.373922 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjk7m\" (UniqueName: \"kubernetes.io/projected/48592971-02b8-495c-8d9e-80956362e01e-kube-api-access-tjk7m\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv\" (UID: \"48592971-02b8-495c-8d9e-80956362e01e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.374000 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvd9b\" (UniqueName: \"kubernetes.io/projected/059b3d3c-de5a-490e-94ff-f15b249de3b3-kube-api-access-fvd9b\") pod \"watcher-operator-controller-manager-5d98cc5575-l5k6p\" (UID: \"059b3d3c-de5a-490e-94ff-f15b249de3b3\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.374309 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c-cert\") pod \"openstack-operator-controller-manager-7996577b9d-7kzr4\" (UID: \"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c\") " pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:13 crc kubenswrapper[4770]: E1004 03:19:13.374405 4770 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 04 03:19:13 crc kubenswrapper[4770]: E1004 03:19:13.374463 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c-cert podName:ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c nodeName:}" failed. No retries permitted until 2025-10-04 03:19:13.874443931 +0000 UTC m=+965.166453643 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c-cert") pod "openstack-operator-controller-manager-7996577b9d-7kzr4" (UID: "ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c") : secret "webhook-server-cert" not found Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.374595 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7w8s\" (UniqueName: \"kubernetes.io/projected/ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c-kube-api-access-s7w8s\") pod \"openstack-operator-controller-manager-7996577b9d-7kzr4\" (UID: \"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c\") " pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.394513 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvd9b\" (UniqueName: \"kubernetes.io/projected/059b3d3c-de5a-490e-94ff-f15b249de3b3-kube-api-access-fvd9b\") pod \"watcher-operator-controller-manager-5d98cc5575-l5k6p\" (UID: \"059b3d3c-de5a-490e-94ff-f15b249de3b3\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.396606 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.402836 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7w8s\" (UniqueName: \"kubernetes.io/projected/ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c-kube-api-access-s7w8s\") pod \"openstack-operator-controller-manager-7996577b9d-7kzr4\" (UID: \"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c\") " pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.477805 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjk7m\" (UniqueName: \"kubernetes.io/projected/48592971-02b8-495c-8d9e-80956362e01e-kube-api-access-tjk7m\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv\" (UID: \"48592971-02b8-495c-8d9e-80956362e01e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.477954 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6231ce27-f748-4e48-ac46-60b1d336e153-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6\" (UID: \"6231ce27-f748-4e48-ac46-60b1d336e153\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.488309 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6231ce27-f748-4e48-ac46-60b1d336e153-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6\" (UID: \"6231ce27-f748-4e48-ac46-60b1d336e153\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.502554 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjk7m\" (UniqueName: \"kubernetes.io/projected/48592971-02b8-495c-8d9e-80956362e01e-kube-api-access-tjk7m\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv\" (UID: \"48592971-02b8-495c-8d9e-80956362e01e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.557031 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.745884 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.888455 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c-cert\") pod \"openstack-operator-controller-manager-7996577b9d-7kzr4\" (UID: \"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c\") " pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:13 crc kubenswrapper[4770]: I1004 03:19:13.900916 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c-cert\") pod \"openstack-operator-controller-manager-7996577b9d-7kzr4\" (UID: \"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c\") " pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.030908 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.158957 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.347337 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" event={"ID":"c0e405f3-b246-4741-8580-248db52f995d","Type":"ContainerStarted","Data":"f2f9953ac30be2da3a9aa82df8f60a4b5cf230e78014ea122a1adfd8e01fe36c"} Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.591592 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz"] Oct 04 03:19:14 crc kubenswrapper[4770]: W1004 03:19:14.594471 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48b4bf81_4d46_4b8c_8a98_2112cb67104c.slice/crio-2c4db538540b50d62080284110f79083a6ce8f084308b6534967bae666b3c53f WatchSource:0}: Error finding container 2c4db538540b50d62080284110f79083a6ce8f084308b6534967bae666b3c53f: Status 404 returned error can't find the container with id 2c4db538540b50d62080284110f79083a6ce8f084308b6534967bae666b3c53f Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.610872 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.615644 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.626123 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.629650 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.634839 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.640226 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.645622 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.649103 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n"] Oct 04 03:19:14 crc kubenswrapper[4770]: W1004 03:19:14.653088 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a0cc679_8960_4bae_9da9_f790f949b5bf.slice/crio-f43c70da3ef7f7fccca0a6f5cf946f292a232db8c577e1667bc7293204a7d3a9 WatchSource:0}: Error finding container f43c70da3ef7f7fccca0a6f5cf946f292a232db8c577e1667bc7293204a7d3a9: Status 404 returned error can't find the container with id f43c70da3ef7f7fccca0a6f5cf946f292a232db8c577e1667bc7293204a7d3a9 Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.653233 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.656853 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.660917 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg"] Oct 04 03:19:14 crc kubenswrapper[4770]: W1004 03:19:14.661525 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1bba079_ef51_40bb_b92d_b03a6e31d457.slice/crio-92a2adf11a8b9e2d4ba3ef5dfc701400b21fd377da3ca3d2771a6825fed83230 WatchSource:0}: Error finding container 92a2adf11a8b9e2d4ba3ef5dfc701400b21fd377da3ca3d2771a6825fed83230: Status 404 returned error can't find the container with id 92a2adf11a8b9e2d4ba3ef5dfc701400b21fd377da3ca3d2771a6825fed83230 Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.702385 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6675647785-zkw64"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.713052 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9"] Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.714197 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hsgxw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-69f59f9d8-fkh9b_openstack-operators(c0126071-94d2-4c21-ac54-ea6de14946ef): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.714609 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rhm7r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-69b956fbf6-rn8fr_openstack-operators(acfbbcdb-767c-4a41-924e-dc60bbd3a5ae): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.738565 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr"] Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.741761 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4wk76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-66f6d6849b-x8ljn_openstack-operators(718d82f1-fa64-4a9d-8780-318b056f44a0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.741854 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-852t7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-c968bb45-bd27z_openstack-operators(2b8720ec-1142-490f-9a6e-9c94e08db137): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.742809 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q7cgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-84788b6bc5-9tpl9_openstack-operators(1fbc59ec-b29e-4843-b34c-43335a590c11): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.742930 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:4cba007c18be1ec9aac2ece7a5ce6444a94afd89f0fb032522811d5bdf5bee73,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dng89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6675647785-zkw64_openstack-operators(d35d4909-8960-4d33-9235-9692bf3a4edd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.773284 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.786878 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.797816 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.806329 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq"] Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.809295 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6"] Oct 04 03:19:14 crc kubenswrapper[4770]: W1004 03:19:14.816322 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod279505bc_0485_4aa9_87ce_930272978c29.slice/crio-1cc6185549245391bd6c5ac7a82b0776e22f4dcd5fb7e267ed83715d2eedff3d WatchSource:0}: Error finding container 1cc6185549245391bd6c5ac7a82b0776e22f4dcd5fb7e267ed83715d2eedff3d: Status 404 returned error can't find the container with id 1cc6185549245391bd6c5ac7a82b0776e22f4dcd5fb7e267ed83715d2eedff3d Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.818054 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:063aae1458289d1090a77c74c2b978b9eb978b0e4062c399f0cb5434a8dd2757,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q944g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7cb48dbc-rwskr_openstack-operators(279505bc-0485-4aa9-87ce-930272978c29): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: W1004 03:19:14.822181 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcfac3b9_2e3b_453e_b434_42159c1f92a9.slice/crio-d0b9f3f025f662ba748a93af59dbfdf923bc535ee450456299c3146988b19ef0 WatchSource:0}: Error finding container d0b9f3f025f662ba748a93af59dbfdf923bc535ee450456299c3146988b19ef0: Status 404 returned error can't find the container with id d0b9f3f025f662ba748a93af59dbfdf923bc535ee450456299c3146988b19ef0 Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.828844 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qtngz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-6c9b57c67-sn2wq_openstack-operators(fcfac3b9-2e3b-453e-b434-42159c1f92a9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: W1004 03:19:14.835445 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6231ce27_f748_4e48_ac46_60b1d336e153.slice/crio-a348d3e9438c8780f7d45f3172c1087241ff2bde1d41856b29084e8d9093440d WatchSource:0}: Error finding container a348d3e9438c8780f7d45f3172c1087241ff2bde1d41856b29084e8d9093440d: Status 404 returned error can't find the container with id a348d3e9438c8780f7d45f3172c1087241ff2bde1d41856b29084e8d9093440d Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.837812 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:d1fad97d2cd602a4f7b6fd6c202464ac117b20e6608c17aa04cadbceb78a498d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:6722a752fb7cbffbae811f6ad6567120fbd4ebbe8c38a83ec2df02850a3276bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:c80fca8178ade7c7eaf9466a74a7cd7e904a7699590ea332acf0ac7bd90e647a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:97cfb1305d59fe73caaaf9d261d08ee16ef3f50e0f3a07fec40fa5c93e5b5190,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:39c22d1455b7278118c6831541db7741712dcc7106a78e8a86ba7e5bcb3a9f23,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:ef02cdf1a68a8cc6ee02f7dcd7e0ca11c828e027069e2e2d6d49c2caa7e6cd70,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:7db347424a8c5998059c5bf84c86a1ef8d582d1ffe39f4887551f2ac85a4915f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:b9a6a9d3bec948e74b8275ada997d8ce394536ad3e2e0fef38ba2f6d643ee560,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:1ee11bb6061765cfedb3c49395f5fb3c0ec537c15e647766fd5f8e9625e587ac,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:0e5405cc4eea290c668815f6ae25b41233c371001037440e0e2752d40aecd438,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:7378049e4fc2b54c20806f9633e831f88d649442facbaf7fa20da005b20da68b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:cb1205166ac37c39e585514884cd7e474b6ec15744290e50c858f2336613532d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:55d94c0ffc3092966f7f5e0539790112ee027faa0e0b7f29972be747dea42e6a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:d61ae6cef3d31df7dfa7375cf5d4975c76ef7ec2e306267f9cfcb212846d15a5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:f90535885474f3b5e262f1ce87c7d3724d8b66f83fc9fbd646c2f0914e1f5597,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:307905acc0883211e04209e4e4692a8bae080e4a3a168bffe4c06ec5f73ebc76,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:f8a2d3c173473e08b57b2741614d8f1ef0f469b07038d85a30687e1b6c6ad32f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:32cae89e17a4562d52b7c28e9f286c4741cef602d042e6bba987c535a17c01cc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:1ed863535bbbadc6a3de00ea5fb4ed5a44b22fa475b0001506327b23b8974f16,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:4ac7d4dac6af6bab407e91d373d93fedd735464e08e72c293eb5ba69210c2e2e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:67e705d98cd50f81c948686b39ecdca6883f0d822e03a697e573111acbc47395,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:36cd868f24a1d5d108841eca240cfd6592792f0ad626addd6cf79a79e638ce62,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:da8363d88d3c5ccdd18903977a44e4ef5c315adbb462c2a8987d4260405f9427,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:cf91f2bb1627b4449e6016194b52182cce38432d0549dcd62adbb0026ccfbfb5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:95875e2d73a4ee08c4f6d092dab04dc167cc3c8b156f29acf0dcde8e220b6672,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:c3bf86219494451daaee9dd376e852a4b9994f301d9bf76648b48e5dfc54625f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:2e3729d3462bc8da3bf09d4176fda3bba9b5ca167201cb108b862dd5fb5a8c67,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:0e78c6c4513b82ee20c1fbf3926ff88e3e0e23305716f89cadd25c885a57dc8f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:2d999c2b47273e04dfbacda96d56cd6d32e8ea1c4e8d6f5c3211e9ff06be69e0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:1505d142ca95c7fd6b9a8eec753423ad6c49c98d71fb5c503e938dec10cecb05,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:7e0e75413d684196317f10b3920f576def3a9a2ffe9e3c24de838562836e983e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:0da1f2f9accb91d893fdb3273c6e66481c8d00b84b262c8cd772a6aaa5ecd1c4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:b2f6e417260d4ae6bf70c05bf2243a0b6cac555cd28080772546761015674191,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:2517c11de03cac2a402dff8575a12f7feec80ad815a83a8a286d8de9d8a5ef9f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:14a280dd80040c348605adaff57e7a66371f8be9cec2d9fcf5101cdae20b2cb7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:af91e545665b439e95fb4bb5d33cc8c5b9e8e38c7cdd35f412e32dc30a4c7c7e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:656a6af048f632a755017d320b68a6e8215179253c9530848e1d6a4f9594fff8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:59a967078b7538e3b0d21b2c23a7d9abe69b75f0dccbeb89777f2f6a91c46170,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:6a0403113cf843d54521055ae34d368f350a79adc66a31da34cfbad7517092ee,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:92c22647ba003cacb23ec491fddfedb346ffabc733e85169bc9ffce4598a35f5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:a4a766bfc1ebcb531717646ef6a7f608c7ee58c883c58c37f1ec80d5062bd672,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:d0a4c0ba8b21503531f6658909a4791db89d34c492b2d0a148e1908d4b0fd835,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:2af885158034339e8d0db2f58d2b27aa822df05b0f7a9c43225e7346b8a0aeeb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:eeb788d6769b1cc1bb421db442cfa457fb575be51440ffe37f39ded03e8b911a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:8a111f93924ba33c7c5baaf8058470208e55eb0b113241539fad9ddb0e4ee916,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:9204104284dfd12ee72833529e8f4244e340e5443cdf2ec5ff25fd5e7b89169c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:352cef8108e0f387a8a1b4d4eea464d094ea5eca7b81543b5e1baeb2d5ecca0d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:374b9dc69979de9ffcec44f060dec37b126d02e12b303177ff1904b4fe79c331,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:014f3c73f78e17359b947bf7b5c38284afa6fa123598f96dc3a9b9d480947827,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:1ae2548f40ab4498166685f922b61946708d1204fd792c09d4256c7a5c86121e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:9ba5c3c71e2f86549334c46e4c411e3a612b8ecc23865c9be55341dcb4c40568,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:26f9f12887eed6bb003c56049af1e56a7da8f8845d78dfaf3d5e9278b5815a30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:29ece1e157dcc25b267e79b7863d35ac453d7751fee005f136beb320a2a888bb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:82a8fbd0df08c8e89109c4aa251f5ffa8055592d416874cdd65f3bff90c06ecf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:5b4c0ddc3f4f06193490b9290007ce0cd098ce0e7ec16c3ccb2759c12d683e51,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:027992dcc115054342aed3fcb65fa94667288d0e443e93bddddc0ceec908da20,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:6e7fc684bf020d3552774ed9a2033634e54e0c946741500f3b93c83a0ede01cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:4a57e6c31fc2727f0cc09d5ee3832779d43404a4237c2328fb2faf9e3a6d0e50,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:432c5410505a3089d1d5ec3f46b39e551970763b7e2f0655f617d9cbad68e138,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:0226892f8e96ba418dd5adcae8d459a5e069b938e7c7abc37335271f9b3b990d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:d6def164f37ae70f9b26399b679048582f4d357fb7a130af69947b63ba68e829,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:5670f9e696b19b76695bea5d4f9c46ac6494d96282f094de1243d8d7a06453b2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:943eee724277e252795909137538a553ef5284c8103ad01b9be7b0138c66d14d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:948b381cd228ded38b6e51b2f297c70f7033a227b565ee8b12490ac48816d4f8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:db16796e99a5906da65eb9068447771f2a11bb3cd6a13cbc2b9c94205ca114a1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:f62e51af159d4420615c3c9045a540af6c3580389e49e48ea9297d748fc09a9b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:9182307b83cce6c08fd9da227d93f1b94e3840ba5c645776d64b7d6e59333ddf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:942c5542e9e5cb8122f7614dd1e8f34badfcc94ed56cc3cf68cfae744392a290,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:45e561fbabeefac4e9f5da933b26bf960576908daf3b09364ac7890b370d0ccc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:d4062383c3b55d604b842067a258b63caf51dcea7d26447a53cd681105626951,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:d142e6e9123b2a4f16d1d65c5d11a132ae9755c2c8cf429ca7ef8c9cd00f4f42,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:b73bcff68993217b4bb909a1fc0191df006be2e4eef48d9bc65a2e3cb0adba0c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:a9f130c7e66a99c8c29b1d8795310ffd4dfa6eb18df3484b87cbcfed9f285406,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:751bc25410e670688b7691763b11a500ed90b8d6dbb84e682cba0db34e743dd4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:5af48acd9260f889cdcbeb2d43cd83aa6a7f3c12b0a9f0d3cedf43e98aed60d6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:3bfbd3e9b524a5152ebd5a70d0412b0f2b6c8b2143b30e49ed4528f651700fdf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:4620884099776ce989a689510bcb41ac167d884292e30d31d4d89d2b08b3c0be,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:65a4ba0d6ebb973b3f0fec8bf2acd2cf99862c70b7f499f8507434184533632d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:0637803c6ceebcf9093b3c8f679f9a5a5be77ea52f530b8c52ba830168433fc2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jd7h6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6_openstack-operators(6231ce27-f748-4e48-ac46-60b1d336e153): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.846883 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4"] Oct 04 03:19:14 crc kubenswrapper[4770]: W1004 03:19:14.862645 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podead820e4_bb4f_4aaa_a35c_fd3f1ac3460c.slice/crio-d206ea784e5ffa87e6f12b983836317675d768df8e08fd8cecd6a58f9dd32821 WatchSource:0}: Error finding container d206ea784e5ffa87e6f12b983836317675d768df8e08fd8cecd6a58f9dd32821: Status 404 returned error can't find the container with id d206ea784e5ffa87e6f12b983836317675d768df8e08fd8cecd6a58f9dd32821 Oct 04 03:19:14 crc kubenswrapper[4770]: I1004 03:19:14.937974 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv"] Oct 04 03:19:14 crc kubenswrapper[4770]: W1004 03:19:14.957572 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48592971_02b8_495c_8d9e_80956362e01e.slice/crio-9dac6528d6c01e3377b3493ddc52cd6fbab4dccc3701172f8c0038c2ca29104c WatchSource:0}: Error finding container 9dac6528d6c01e3377b3493ddc52cd6fbab4dccc3701172f8c0038c2ca29104c: Status 404 returned error can't find the container with id 9dac6528d6c01e3377b3493ddc52cd6fbab4dccc3701172f8c0038c2ca29104c Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.972272 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tjk7m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv_openstack-operators(48592971-02b8-495c-8d9e-80956362e01e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:19:14 crc kubenswrapper[4770]: E1004 03:19:14.974169 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" podUID="48592971-02b8-495c-8d9e-80956362e01e" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:14.999973 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" podUID="acfbbcdb-767c-4a41-924e-dc60bbd3a5ae" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.019143 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" podUID="c0126071-94d2-4c21-ac54-ea6de14946ef" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.049445 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" podUID="1fbc59ec-b29e-4843-b34c-43335a590c11" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.116300 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" podUID="fcfac3b9-2e3b-453e-b434-42159c1f92a9" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.145762 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" podUID="718d82f1-fa64-4a9d-8780-318b056f44a0" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.214299 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" podUID="6231ce27-f748-4e48-ac46-60b1d336e153" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.240279 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" podUID="d35d4909-8960-4d33-9235-9692bf3a4edd" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.245839 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" podUID="279505bc-0485-4aa9-87ce-930272978c29" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.294044 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" podUID="2b8720ec-1142-490f-9a6e-9c94e08db137" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.438341 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" event={"ID":"aab021e2-b955-4efa-8668-1444f506ca5d","Type":"ContainerStarted","Data":"20ad6dd780d51328379dc7f553c112c6904a6611e1b1e67db7fafa0e85fa9f35"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.441670 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" event={"ID":"279505bc-0485-4aa9-87ce-930272978c29","Type":"ContainerStarted","Data":"0df8d59ca178a000c3b98868a3c9e2d3241c30cdcd8c4ae7b0067bc078459722"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.441745 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" event={"ID":"279505bc-0485-4aa9-87ce-930272978c29","Type":"ContainerStarted","Data":"1cc6185549245391bd6c5ac7a82b0776e22f4dcd5fb7e267ed83715d2eedff3d"} Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.462828 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:063aae1458289d1090a77c74c2b978b9eb978b0e4062c399f0cb5434a8dd2757\\\"\"" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" podUID="279505bc-0485-4aa9-87ce-930272978c29" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.474208 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" event={"ID":"48592971-02b8-495c-8d9e-80956362e01e","Type":"ContainerStarted","Data":"9dac6528d6c01e3377b3493ddc52cd6fbab4dccc3701172f8c0038c2ca29104c"} Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.479362 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" podUID="48592971-02b8-495c-8d9e-80956362e01e" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.503357 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" event={"ID":"72f16f95-3791-461e-8d91-d89005030236","Type":"ContainerStarted","Data":"fc6cceadf3e52a8e3c5e27f0bd11aba59efa99736efca8e2103ecd5a64f48552"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.534497 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" event={"ID":"d35d4909-8960-4d33-9235-9692bf3a4edd","Type":"ContainerStarted","Data":"8103cd3077ff7641f9a8db3046d9076a36ae5e9f80c0f5c62a72329594e47fbb"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.534567 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" event={"ID":"d35d4909-8960-4d33-9235-9692bf3a4edd","Type":"ContainerStarted","Data":"0aa86e9fdc586563f320a43773e418f7bae39438dc20372e50cf22ce6eb4d17e"} Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.537723 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:4cba007c18be1ec9aac2ece7a5ce6444a94afd89f0fb032522811d5bdf5bee73\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" podUID="d35d4909-8960-4d33-9235-9692bf3a4edd" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.541860 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" event={"ID":"c0126071-94d2-4c21-ac54-ea6de14946ef","Type":"ContainerStarted","Data":"cadc63b5b00ac72ec77a34874726207b5e4afeb84cef63c4a9a29c06573127d6"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.541923 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" event={"ID":"c0126071-94d2-4c21-ac54-ea6de14946ef","Type":"ContainerStarted","Data":"23196d4df8d79ae08fa5b71c2f40cb27b9df19fb0e8649ea5990b5b6047c080e"} Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.559796 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" podUID="c0126071-94d2-4c21-ac54-ea6de14946ef" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.568571 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" event={"ID":"059b3d3c-de5a-490e-94ff-f15b249de3b3","Type":"ContainerStarted","Data":"8194fbea3f5b22487d3d6209d31e7a5aa16f332fcfef3f84c3c07dc1d79f8d28"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.595401 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" event={"ID":"64c06e35-b80f-4889-a680-36c627c99a06","Type":"ContainerStarted","Data":"b26a169eb05714d7015ca80f34ab7d1729076aa5e2f29faccab80117027956b0"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.598028 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" event={"ID":"e26cfeef-31d4-452e-bbf6-f54ab0d3b331","Type":"ContainerStarted","Data":"6c9e2bea023be02bd85a23638fbf4fa13bfd17164ac7aa6691f5541ddeda96cb"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.600693 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" event={"ID":"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c","Type":"ContainerStarted","Data":"307c2f51d1d29c6769045fe244fb8edad6155f1b533cf3a0f80d8e5de2b105b6"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.600755 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" event={"ID":"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c","Type":"ContainerStarted","Data":"d206ea784e5ffa87e6f12b983836317675d768df8e08fd8cecd6a58f9dd32821"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.602506 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.617627 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" event={"ID":"718d82f1-fa64-4a9d-8780-318b056f44a0","Type":"ContainerStarted","Data":"f5b03dddb58693092f1510f92b7ac4569f435c604bee8f773a64a3a4142b96c1"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.617685 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" event={"ID":"718d82f1-fa64-4a9d-8780-318b056f44a0","Type":"ContainerStarted","Data":"90e45932f50abf3b2a9c6e0897390b18c88202d29e46b2afc33f50e6a568f69a"} Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.626402 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" podUID="718d82f1-fa64-4a9d-8780-318b056f44a0" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.632359 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" event={"ID":"09622ed2-b879-4add-9df7-0fb77dfa85cb","Type":"ContainerStarted","Data":"4e0a27306816f2452487e3a303d96bd7f8eef4cae3a0943790a1a9c231a3ea86"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.661305 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" event={"ID":"48b4bf81-4d46-4b8c-8a98-2112cb67104c","Type":"ContainerStarted","Data":"2c4db538540b50d62080284110f79083a6ce8f084308b6534967bae666b3c53f"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.663416 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" event={"ID":"6231ce27-f748-4e48-ac46-60b1d336e153","Type":"ContainerStarted","Data":"c84ea32948487cd15a39d7b66a999c35192b56adf1db3d7e433c55144bf84e47"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.663443 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" event={"ID":"6231ce27-f748-4e48-ac46-60b1d336e153","Type":"ContainerStarted","Data":"a348d3e9438c8780f7d45f3172c1087241ff2bde1d41856b29084e8d9093440d"} Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.678182 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" podUID="6231ce27-f748-4e48-ac46-60b1d336e153" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.685752 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" podStartSLOduration=2.685713785 podStartE2EDuration="2.685713785s" podCreationTimestamp="2025-10-04 03:19:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:19:15.668980328 +0000 UTC m=+966.960990040" watchObservedRunningTime="2025-10-04 03:19:15.685713785 +0000 UTC m=+966.977723498" Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.707203 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" podUID="acfbbcdb-767c-4a41-924e-dc60bbd3a5ae" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.712879 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" event={"ID":"acfbbcdb-767c-4a41-924e-dc60bbd3a5ae","Type":"ContainerStarted","Data":"1006b0aa51c39b4ee4895539f19896990f9b710de2ffea44e85a5add0caa3d65"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.712950 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" event={"ID":"acfbbcdb-767c-4a41-924e-dc60bbd3a5ae","Type":"ContainerStarted","Data":"4fdd07c44562ecf7ab41dc1b738098ae58c020272dfa1937a052eb9bdb8a3b90"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.712966 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" event={"ID":"34a19c5f-604f-4225-be1d-c66d51b0919d","Type":"ContainerStarted","Data":"05843f8ac079f81bcb1b5992d990053028c2836f22d0898f6b63474bdd3f7519"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.713280 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" event={"ID":"7a0cc679-8960-4bae-9da9-f790f949b5bf","Type":"ContainerStarted","Data":"f43c70da3ef7f7fccca0a6f5cf946f292a232db8c577e1667bc7293204a7d3a9"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.727944 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" event={"ID":"909aa645-88e5-4e67-835e-34854c2f5a74","Type":"ContainerStarted","Data":"cf654287de97b227994b4ca5dcc48130678d71e9f1115ae92b9fb5816f781df3"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.741420 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" event={"ID":"c1bba079-ef51-40bb-b92d-b03a6e31d457","Type":"ContainerStarted","Data":"92a2adf11a8b9e2d4ba3ef5dfc701400b21fd377da3ca3d2771a6825fed83230"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.746765 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" event={"ID":"fcfac3b9-2e3b-453e-b434-42159c1f92a9","Type":"ContainerStarted","Data":"fee495cd40fc557ae35ea84f039b7d8b10ed2a39f49afed4952ed5d5471d1ddf"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.746801 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" event={"ID":"fcfac3b9-2e3b-453e-b434-42159c1f92a9","Type":"ContainerStarted","Data":"d0b9f3f025f662ba748a93af59dbfdf923bc535ee450456299c3146988b19ef0"} Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.752073 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" podUID="fcfac3b9-2e3b-453e-b434-42159c1f92a9" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.809193 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" event={"ID":"2b8720ec-1142-490f-9a6e-9c94e08db137","Type":"ContainerStarted","Data":"76d5f14631d0b1574343fdf93d69717c8833bc30eaf4addf546b1f557654c0cf"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.809245 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" event={"ID":"2b8720ec-1142-490f-9a6e-9c94e08db137","Type":"ContainerStarted","Data":"0bcfca3cd60371e48316c7cde306283f233fc4cb57ae55a4ac9c16a00b12e8a2"} Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.810770 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" podUID="2b8720ec-1142-490f-9a6e-9c94e08db137" Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.821945 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" event={"ID":"1fbc59ec-b29e-4843-b34c-43335a590c11","Type":"ContainerStarted","Data":"270614442e0b86ceb103807fe077bf96a727a294dbaf2c6c436889400f1f412e"} Oct 04 03:19:15 crc kubenswrapper[4770]: I1004 03:19:15.822025 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" event={"ID":"1fbc59ec-b29e-4843-b34c-43335a590c11","Type":"ContainerStarted","Data":"4079e744ee981314089d9c28739fc250aae30e28e8353557e347d3d2582c0b8a"} Oct 04 03:19:15 crc kubenswrapper[4770]: E1004 03:19:15.830565 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" podUID="1fbc59ec-b29e-4843-b34c-43335a590c11" Oct 04 03:19:16 crc kubenswrapper[4770]: I1004 03:19:16.847901 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" event={"ID":"ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c","Type":"ContainerStarted","Data":"a5b6842832969247705d6bcdca9637b68265707b9f922383a027c692c8244679"} Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.851520 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" podUID="c0126071-94d2-4c21-ac54-ea6de14946ef" Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.851688 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" podUID="2b8720ec-1142-490f-9a6e-9c94e08db137" Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.851797 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:063aae1458289d1090a77c74c2b978b9eb978b0e4062c399f0cb5434a8dd2757\\\"\"" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" podUID="279505bc-0485-4aa9-87ce-930272978c29" Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.855925 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" podUID="6231ce27-f748-4e48-ac46-60b1d336e153" Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.855992 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:4cba007c18be1ec9aac2ece7a5ce6444a94afd89f0fb032522811d5bdf5bee73\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" podUID="d35d4909-8960-4d33-9235-9692bf3a4edd" Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.856114 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" podUID="fcfac3b9-2e3b-453e-b434-42159c1f92a9" Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.856111 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" podUID="48592971-02b8-495c-8d9e-80956362e01e" Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.856199 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" podUID="acfbbcdb-767c-4a41-924e-dc60bbd3a5ae" Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.856357 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" podUID="718d82f1-fa64-4a9d-8780-318b056f44a0" Oct 04 03:19:16 crc kubenswrapper[4770]: E1004 03:19:16.856625 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" podUID="1fbc59ec-b29e-4843-b34c-43335a590c11" Oct 04 03:19:22 crc kubenswrapper[4770]: I1004 03:19:22.929067 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" event={"ID":"e26cfeef-31d4-452e-bbf6-f54ab0d3b331","Type":"ContainerStarted","Data":"47b93f521229efa9c65dd51391800365afd3596bbc73b9d05e998d494614939e"} Oct 04 03:19:22 crc kubenswrapper[4770]: I1004 03:19:22.938678 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" event={"ID":"34a19c5f-604f-4225-be1d-c66d51b0919d","Type":"ContainerStarted","Data":"e93732d936d666beadeda00440e2afe3c085a49fe94fea1e8559611f573127a3"} Oct 04 03:19:22 crc kubenswrapper[4770]: I1004 03:19:22.948233 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" event={"ID":"7a0cc679-8960-4bae-9da9-f790f949b5bf","Type":"ContainerStarted","Data":"00b56a2d495ba16254fb6879aa5124784d9c3dd8b636ee3bfc22f8dbca649ed5"} Oct 04 03:19:22 crc kubenswrapper[4770]: I1004 03:19:22.989112 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" event={"ID":"059b3d3c-de5a-490e-94ff-f15b249de3b3","Type":"ContainerStarted","Data":"143f8ea3a8dd4b421892ded8db5a587a97ee269647237f8c2a20b986f8fd1df9"} Oct 04 03:19:22 crc kubenswrapper[4770]: I1004 03:19:22.997241 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" event={"ID":"48b4bf81-4d46-4b8c-8a98-2112cb67104c","Type":"ContainerStarted","Data":"567577f8a0992d034cf2d91c3ca21dd52c938d006b4ed7e4b5734119cf4e54b4"} Oct 04 03:19:23 crc kubenswrapper[4770]: I1004 03:19:23.004565 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" event={"ID":"909aa645-88e5-4e67-835e-34854c2f5a74","Type":"ContainerStarted","Data":"6cff93ceed4462c1670e3d36d897f20d05abdda0bc746126e232c52265a481ff"} Oct 04 03:19:23 crc kubenswrapper[4770]: I1004 03:19:23.006277 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" event={"ID":"72f16f95-3791-461e-8d91-d89005030236","Type":"ContainerStarted","Data":"96c8bdc6c9bd359b606a29af0edddd4d65917b7d309e38124f3e07957e94acd2"} Oct 04 03:19:23 crc kubenswrapper[4770]: I1004 03:19:23.024342 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" event={"ID":"c0e405f3-b246-4741-8580-248db52f995d","Type":"ContainerStarted","Data":"6ea35ba5eac48bd5e3998fc32bf00a9336c3424fdc17273f7cc86c4f40348f71"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.033384 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" event={"ID":"7a0cc679-8960-4bae-9da9-f790f949b5bf","Type":"ContainerStarted","Data":"08969a21ec729bd3242996cdee7b21fdc7e844a50e1bac873325ecff14bdd3a1"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.033827 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.035164 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" event={"ID":"909aa645-88e5-4e67-835e-34854c2f5a74","Type":"ContainerStarted","Data":"49bb8931fea380ac3f28dd45c53e5a03cd53fcd0c0e238179cd9f28186fc86f0"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.035355 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.037102 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" event={"ID":"72f16f95-3791-461e-8d91-d89005030236","Type":"ContainerStarted","Data":"1001caf956a03588d502a590d512439a1b0d288fcb9cb2a0c609187c346d6c43"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.037181 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.038797 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" event={"ID":"aab021e2-b955-4efa-8668-1444f506ca5d","Type":"ContainerStarted","Data":"b149e057b24e490d3f4beb67084f3ef98b437ec17468d999bd7eb0cc13189cf1"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.038838 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" event={"ID":"aab021e2-b955-4efa-8668-1444f506ca5d","Type":"ContainerStarted","Data":"a79260683c50b59d868190c4c7b9d45881140a611c2c6456a0978d9b29a8c6ea"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.038909 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.040362 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" event={"ID":"09622ed2-b879-4add-9df7-0fb77dfa85cb","Type":"ContainerStarted","Data":"5f15168130bce7e1d3eee71681e9befa109e0d59c374b60d53d03015e797fc6f"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.040385 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" event={"ID":"09622ed2-b879-4add-9df7-0fb77dfa85cb","Type":"ContainerStarted","Data":"e638a087fd2eeb3b7622fe97a5baac9f16c1d088b92f7bf5c0f8a2580974db1b"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.040953 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.045228 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7996577b9d-7kzr4" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.049593 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" event={"ID":"64c06e35-b80f-4889-a680-36c627c99a06","Type":"ContainerStarted","Data":"18f7d6185910853ba0dd65d6ab82b1ad3375349bb56447a13178a9c296cc4443"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.049644 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" event={"ID":"64c06e35-b80f-4889-a680-36c627c99a06","Type":"ContainerStarted","Data":"8950cc49092fa0cb88d3e7e8a55128a6ef8721e1a8176b90eb0d9ed7e1e83bdf"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.049664 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.051404 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" event={"ID":"48b4bf81-4d46-4b8c-8a98-2112cb67104c","Type":"ContainerStarted","Data":"26b4426ccf553f2ac666e96f16c3cc91f806a205a3ee6c833b6d8d5422179700"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.051775 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.053433 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" event={"ID":"e26cfeef-31d4-452e-bbf6-f54ab0d3b331","Type":"ContainerStarted","Data":"f608cb7af3055251d8e9741f49f6056cb41a395b27996776e4d7c226510efe5d"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.053826 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.056840 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" event={"ID":"34a19c5f-604f-4225-be1d-c66d51b0919d","Type":"ContainerStarted","Data":"60707e1f03ea1a3681220c37740474b699e5e703bb260ca5429f0b387cb546c2"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.057211 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.059443 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" event={"ID":"c0e405f3-b246-4741-8580-248db52f995d","Type":"ContainerStarted","Data":"0bc63e4d0fb509840b7a8fdf444900cfffeefddf22fb2aca3ddfed1406381c94"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.059805 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.061330 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" event={"ID":"c1bba079-ef51-40bb-b92d-b03a6e31d457","Type":"ContainerStarted","Data":"b0760977039530b94c91977d87244cb718887e61e4a4d5c0aab04ec50d35b0fa"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.061359 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" event={"ID":"c1bba079-ef51-40bb-b92d-b03a6e31d457","Type":"ContainerStarted","Data":"4d9c633999412052fba27723594e4133caa1a98204f872e941737c4f1945a96d"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.061825 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.063636 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" event={"ID":"059b3d3c-de5a-490e-94ff-f15b249de3b3","Type":"ContainerStarted","Data":"ef68924d6db8886131c2d5a5f9b44c6b583c719f2499fd8f61f95997ef5e958a"} Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.064050 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.071153 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" podStartSLOduration=4.46039486 podStartE2EDuration="12.07113133s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.662542012 +0000 UTC m=+965.954551724" lastFinishedPulling="2025-10-04 03:19:22.273278442 +0000 UTC m=+973.565288194" observedRunningTime="2025-10-04 03:19:24.065167283 +0000 UTC m=+975.357177005" watchObservedRunningTime="2025-10-04 03:19:24.07113133 +0000 UTC m=+975.363141042" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.091135 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" podStartSLOduration=4.491300629 podStartE2EDuration="12.091114112s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.673497259 +0000 UTC m=+965.965506971" lastFinishedPulling="2025-10-04 03:19:22.273310742 +0000 UTC m=+973.565320454" observedRunningTime="2025-10-04 03:19:24.086177933 +0000 UTC m=+975.378187665" watchObservedRunningTime="2025-10-04 03:19:24.091114112 +0000 UTC m=+975.383123824" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.106006 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" podStartSLOduration=4.012683113 podStartE2EDuration="12.105985751s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.176697178 +0000 UTC m=+965.468706890" lastFinishedPulling="2025-10-04 03:19:22.269999796 +0000 UTC m=+973.562009528" observedRunningTime="2025-10-04 03:19:24.104529123 +0000 UTC m=+975.396538845" watchObservedRunningTime="2025-10-04 03:19:24.105985751 +0000 UTC m=+975.397995463" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.121398 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" podStartSLOduration=4.517998996 podStartE2EDuration="12.121373204s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.708493924 +0000 UTC m=+966.000503636" lastFinishedPulling="2025-10-04 03:19:22.311868112 +0000 UTC m=+973.603877844" observedRunningTime="2025-10-04 03:19:24.120153863 +0000 UTC m=+975.412163575" watchObservedRunningTime="2025-10-04 03:19:24.121373204 +0000 UTC m=+975.413382936" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.194643 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" podStartSLOduration=4.647118315 podStartE2EDuration="12.194629332s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.708277857 +0000 UTC m=+966.000287569" lastFinishedPulling="2025-10-04 03:19:22.255788844 +0000 UTC m=+973.547798586" observedRunningTime="2025-10-04 03:19:24.191235513 +0000 UTC m=+975.483245225" watchObservedRunningTime="2025-10-04 03:19:24.194629332 +0000 UTC m=+975.486639044" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.245678 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" podStartSLOduration=4.629715841 podStartE2EDuration="12.245661477s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.657285084 +0000 UTC m=+965.949294796" lastFinishedPulling="2025-10-04 03:19:22.2732307 +0000 UTC m=+973.565240432" observedRunningTime="2025-10-04 03:19:24.244830386 +0000 UTC m=+975.536840098" watchObservedRunningTime="2025-10-04 03:19:24.245661477 +0000 UTC m=+975.537671189" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.250161 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" podStartSLOduration=4.657336003 podStartE2EDuration="12.250149165s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.680553123 +0000 UTC m=+965.972562835" lastFinishedPulling="2025-10-04 03:19:22.273366275 +0000 UTC m=+973.565375997" observedRunningTime="2025-10-04 03:19:24.225851359 +0000 UTC m=+975.517861071" watchObservedRunningTime="2025-10-04 03:19:24.250149165 +0000 UTC m=+975.542158877" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.266043 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" podStartSLOduration=4.804581444 podStartE2EDuration="12.266028501s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.813769546 +0000 UTC m=+966.105779258" lastFinishedPulling="2025-10-04 03:19:22.275216583 +0000 UTC m=+973.567226315" observedRunningTime="2025-10-04 03:19:24.261681867 +0000 UTC m=+975.553691579" watchObservedRunningTime="2025-10-04 03:19:24.266028501 +0000 UTC m=+975.558038213" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.289452 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" podStartSLOduration=4.676146707 podStartE2EDuration="12.289431713s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.659972685 +0000 UTC m=+965.951982387" lastFinishedPulling="2025-10-04 03:19:22.273257661 +0000 UTC m=+973.565267393" observedRunningTime="2025-10-04 03:19:24.284665038 +0000 UTC m=+975.576674750" watchObservedRunningTime="2025-10-04 03:19:24.289431713 +0000 UTC m=+975.581441435" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.317033 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" podStartSLOduration=4.739818623 podStartE2EDuration="12.316997604s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.666514916 +0000 UTC m=+965.958524628" lastFinishedPulling="2025-10-04 03:19:22.243693857 +0000 UTC m=+973.535703609" observedRunningTime="2025-10-04 03:19:24.307452834 +0000 UTC m=+975.599462546" watchObservedRunningTime="2025-10-04 03:19:24.316997604 +0000 UTC m=+975.609007306" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.326947 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" podStartSLOduration=4.641204833 podStartE2EDuration="12.326931724s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.596995058 +0000 UTC m=+965.889004770" lastFinishedPulling="2025-10-04 03:19:22.282721949 +0000 UTC m=+973.574731661" observedRunningTime="2025-10-04 03:19:24.323366202 +0000 UTC m=+975.615375904" watchObservedRunningTime="2025-10-04 03:19:24.326931724 +0000 UTC m=+975.618941436" Oct 04 03:19:24 crc kubenswrapper[4770]: I1004 03:19:24.339249 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" podStartSLOduration=4.732485982 podStartE2EDuration="12.339228447s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.666561007 +0000 UTC m=+965.958570719" lastFinishedPulling="2025-10-04 03:19:22.273303452 +0000 UTC m=+973.565313184" observedRunningTime="2025-10-04 03:19:24.337736667 +0000 UTC m=+975.629746389" watchObservedRunningTime="2025-10-04 03:19:24.339228447 +0000 UTC m=+975.631238159" Oct 04 03:19:27 crc kubenswrapper[4770]: I1004 03:19:27.677111 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:19:30 crc kubenswrapper[4770]: I1004 03:19:30.120747 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" event={"ID":"6231ce27-f748-4e48-ac46-60b1d336e153","Type":"ContainerStarted","Data":"0725d25decdd71f4b30aa51d14f757b9adc29654d1ab66ad0af29c0e1a58e76c"} Oct 04 03:19:30 crc kubenswrapper[4770]: I1004 03:19:30.122958 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:19:32 crc kubenswrapper[4770]: I1004 03:19:32.578999 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-m269q" Oct 04 03:19:32 crc kubenswrapper[4770]: I1004 03:19:32.592325 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-nq6dg" Oct 04 03:19:32 crc kubenswrapper[4770]: I1004 03:19:32.605975 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" podStartSLOduration=6.095346838 podStartE2EDuration="20.605943451s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.837328592 +0000 UTC m=+966.129338304" lastFinishedPulling="2025-10-04 03:19:29.347925195 +0000 UTC m=+980.639934917" observedRunningTime="2025-10-04 03:19:30.149511826 +0000 UTC m=+981.441521558" watchObservedRunningTime="2025-10-04 03:19:32.605943451 +0000 UTC m=+983.897953163" Oct 04 03:19:32 crc kubenswrapper[4770]: I1004 03:19:32.629106 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-qj4zt" Oct 04 03:19:32 crc kubenswrapper[4770]: I1004 03:19:32.638325 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-dzkch" Oct 04 03:19:32 crc kubenswrapper[4770]: I1004 03:19:32.677037 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-bcfrt" Oct 04 03:19:32 crc kubenswrapper[4770]: I1004 03:19:32.857782 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-56p4n" Oct 04 03:19:32 crc kubenswrapper[4770]: I1004 03:19:32.872067 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-95452" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.036277 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-67q4w" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.172127 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" event={"ID":"48592971-02b8-495c-8d9e-80956362e01e","Type":"ContainerStarted","Data":"f40bc037a351b385e1913f0202882e37d2edec9961f827a60c84ba4b55aedfbd"} Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.191907 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" event={"ID":"c0126071-94d2-4c21-ac54-ea6de14946ef","Type":"ContainerStarted","Data":"6f706d1fe95879136547441e6e41a044a6c25ab025afd27f54821299804b0bd6"} Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.192678 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.201049 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" event={"ID":"279505bc-0485-4aa9-87ce-930272978c29","Type":"ContainerStarted","Data":"528d5a6a400ac92a59e4db9ee4bf396d9add6e73cefa10e14fcb73d131bef43c"} Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.201680 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.203419 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" event={"ID":"fcfac3b9-2e3b-453e-b434-42159c1f92a9","Type":"ContainerStarted","Data":"2ce0a62f5b30be0ae3f6dc999d02ed55889c6bfe55b35d8e920fb986f4f5b69d"} Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.203836 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.207397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" event={"ID":"2b8720ec-1142-490f-9a6e-9c94e08db137","Type":"ContainerStarted","Data":"8c2dd6f569c0a7bb6a79cf052c4a8e381bf24ade6c14cb03e765e88dac31090d"} Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.208046 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.221676 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv" podStartSLOduration=2.718569755 podStartE2EDuration="20.221657657s" podCreationTimestamp="2025-10-04 03:19:13 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.972090736 +0000 UTC m=+966.264100448" lastFinishedPulling="2025-10-04 03:19:32.475178648 +0000 UTC m=+983.767188350" observedRunningTime="2025-10-04 03:19:33.21833968 +0000 UTC m=+984.510349392" watchObservedRunningTime="2025-10-04 03:19:33.221657657 +0000 UTC m=+984.513667379" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.236835 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" podStartSLOduration=3.3754811 podStartE2EDuration="21.236816494s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.741550137 +0000 UTC m=+966.033559849" lastFinishedPulling="2025-10-04 03:19:32.602885541 +0000 UTC m=+983.894895243" observedRunningTime="2025-10-04 03:19:33.235877749 +0000 UTC m=+984.527887451" watchObservedRunningTime="2025-10-04 03:19:33.236816494 +0000 UTC m=+984.528826216" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.272502 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" podStartSLOduration=3.6244035820000002 podStartE2EDuration="21.272470206s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.828709066 +0000 UTC m=+966.120718778" lastFinishedPulling="2025-10-04 03:19:32.47677569 +0000 UTC m=+983.768785402" observedRunningTime="2025-10-04 03:19:33.268668698 +0000 UTC m=+984.560678410" watchObservedRunningTime="2025-10-04 03:19:33.272470206 +0000 UTC m=+984.564479928" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.287508 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" podStartSLOduration=3.525129617 podStartE2EDuration="21.28749102s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.713953935 +0000 UTC m=+966.005963637" lastFinishedPulling="2025-10-04 03:19:32.476315328 +0000 UTC m=+983.768325040" observedRunningTime="2025-10-04 03:19:33.284964174 +0000 UTC m=+984.576973886" watchObservedRunningTime="2025-10-04 03:19:33.28749102 +0000 UTC m=+984.579500732" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.289777 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-76d5577b-8d8z6" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.316215 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" podStartSLOduration=3.657748396 podStartE2EDuration="21.316196311s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.817778021 +0000 UTC m=+966.109787733" lastFinishedPulling="2025-10-04 03:19:32.476225936 +0000 UTC m=+983.768235648" observedRunningTime="2025-10-04 03:19:33.307536814 +0000 UTC m=+984.599546536" watchObservedRunningTime="2025-10-04 03:19:33.316196311 +0000 UTC m=+984.608206023" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.351967 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-26xxz" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.375759 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-mhcm4" Oct 04 03:19:33 crc kubenswrapper[4770]: I1004 03:19:33.400895 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-l5k6p" Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.229352 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" event={"ID":"d35d4909-8960-4d33-9235-9692bf3a4edd","Type":"ContainerStarted","Data":"d471814309890f76ab04577ae9ab960c8df1db784acc215a1dbed0f6dbb7f7f0"} Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.229888 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.231148 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" event={"ID":"718d82f1-fa64-4a9d-8780-318b056f44a0","Type":"ContainerStarted","Data":"8678e9b659fc4b3af83ab55e6d86dd602897bf4576ee90297c7ad5b615ffaaed"} Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.231383 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.232899 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" event={"ID":"1fbc59ec-b29e-4843-b34c-43335a590c11","Type":"ContainerStarted","Data":"e355840b744b823c22ebb5b042da4287e0993076ccfdfc7ef4c8092ee1f15f95"} Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.233111 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.234734 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" event={"ID":"acfbbcdb-767c-4a41-924e-dc60bbd3a5ae","Type":"ContainerStarted","Data":"cc11ed4b50264095a349905d08f0f498c0f91789c35f29478cb7c448b8edd0ec"} Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.234887 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.267307 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" podStartSLOduration=3.395783391 podStartE2EDuration="24.267283173s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.742861632 +0000 UTC m=+966.034871344" lastFinishedPulling="2025-10-04 03:19:35.614361374 +0000 UTC m=+986.906371126" observedRunningTime="2025-10-04 03:19:36.251448359 +0000 UTC m=+987.543458071" watchObservedRunningTime="2025-10-04 03:19:36.267283173 +0000 UTC m=+987.559292885" Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.273107 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" podStartSLOduration=3.400139205 podStartE2EDuration="24.273089305s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.742680157 +0000 UTC m=+966.034689879" lastFinishedPulling="2025-10-04 03:19:35.615630267 +0000 UTC m=+986.907639979" observedRunningTime="2025-10-04 03:19:36.266600296 +0000 UTC m=+987.558610008" watchObservedRunningTime="2025-10-04 03:19:36.273089305 +0000 UTC m=+987.565099017" Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.286595 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" podStartSLOduration=3.409223583 podStartE2EDuration="24.286577948s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.741598869 +0000 UTC m=+966.033609131" lastFinishedPulling="2025-10-04 03:19:35.618953784 +0000 UTC m=+986.910963496" observedRunningTime="2025-10-04 03:19:36.283875147 +0000 UTC m=+987.575884859" watchObservedRunningTime="2025-10-04 03:19:36.286577948 +0000 UTC m=+987.578587650" Oct 04 03:19:36 crc kubenswrapper[4770]: I1004 03:19:36.301341 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" podStartSLOduration=3.39151796 podStartE2EDuration="24.301324994s" podCreationTimestamp="2025-10-04 03:19:12 +0000 UTC" firstStartedPulling="2025-10-04 03:19:14.714537921 +0000 UTC m=+966.006547633" lastFinishedPulling="2025-10-04 03:19:35.624344955 +0000 UTC m=+986.916354667" observedRunningTime="2025-10-04 03:19:36.299625479 +0000 UTC m=+987.591635191" watchObservedRunningTime="2025-10-04 03:19:36.301324994 +0000 UTC m=+987.593334696" Oct 04 03:19:42 crc kubenswrapper[4770]: I1004 03:19:42.771131 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6675647785-zkw64" Oct 04 03:19:42 crc kubenswrapper[4770]: I1004 03:19:42.836309 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-9tpl9" Oct 04 03:19:43 crc kubenswrapper[4770]: I1004 03:19:43.029199 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-rwskr" Oct 04 03:19:43 crc kubenswrapper[4770]: I1004 03:19:43.053993 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-rn8fr" Oct 04 03:19:43 crc kubenswrapper[4770]: I1004 03:19:43.078558 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-fkh9b" Oct 04 03:19:43 crc kubenswrapper[4770]: I1004 03:19:43.106924 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-sn2wq" Oct 04 03:19:43 crc kubenswrapper[4770]: I1004 03:19:43.299047 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-bd27z" Oct 04 03:19:43 crc kubenswrapper[4770]: I1004 03:19:43.335189 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-x8ljn" Oct 04 03:19:43 crc kubenswrapper[4770]: I1004 03:19:43.568260 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6" Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.875737 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-v2tnx"] Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.877779 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.880484 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.880800 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.880919 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.881047 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-mpcdx" Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.884577 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-v2tnx"] Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.980648 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-gtdlx"] Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.982129 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.987860 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 04 03:20:00 crc kubenswrapper[4770]: I1004 03:20:00.994429 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-gtdlx"] Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.027519 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-config\") pod \"dnsmasq-dns-546d69f86c-v2tnx\" (UID: \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\") " pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.027719 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp54d\" (UniqueName: \"kubernetes.io/projected/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-kube-api-access-cp54d\") pod \"dnsmasq-dns-546d69f86c-v2tnx\" (UID: \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\") " pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.129202 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp54d\" (UniqueName: \"kubernetes.io/projected/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-kube-api-access-cp54d\") pod \"dnsmasq-dns-546d69f86c-v2tnx\" (UID: \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\") " pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.129265 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-dns-svc\") pod \"dnsmasq-dns-7f9579fb85-gtdlx\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.129310 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-config\") pod \"dnsmasq-dns-546d69f86c-v2tnx\" (UID: \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\") " pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.129338 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-config\") pod \"dnsmasq-dns-7f9579fb85-gtdlx\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.129376 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7lkg\" (UniqueName: \"kubernetes.io/projected/39554a25-d378-4002-9359-f07a23f8e621-kube-api-access-t7lkg\") pod \"dnsmasq-dns-7f9579fb85-gtdlx\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.130592 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-config\") pod \"dnsmasq-dns-546d69f86c-v2tnx\" (UID: \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\") " pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.155933 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp54d\" (UniqueName: \"kubernetes.io/projected/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-kube-api-access-cp54d\") pod \"dnsmasq-dns-546d69f86c-v2tnx\" (UID: \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\") " pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.205103 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.230570 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-dns-svc\") pod \"dnsmasq-dns-7f9579fb85-gtdlx\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.230705 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-config\") pod \"dnsmasq-dns-7f9579fb85-gtdlx\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.230778 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7lkg\" (UniqueName: \"kubernetes.io/projected/39554a25-d378-4002-9359-f07a23f8e621-kube-api-access-t7lkg\") pod \"dnsmasq-dns-7f9579fb85-gtdlx\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.231778 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-dns-svc\") pod \"dnsmasq-dns-7f9579fb85-gtdlx\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.231817 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-config\") pod \"dnsmasq-dns-7f9579fb85-gtdlx\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.264695 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7lkg\" (UniqueName: \"kubernetes.io/projected/39554a25-d378-4002-9359-f07a23f8e621-kube-api-access-t7lkg\") pod \"dnsmasq-dns-7f9579fb85-gtdlx\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.304047 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.683747 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-v2tnx"] Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.780239 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-gtdlx"] Oct 04 03:20:01 crc kubenswrapper[4770]: W1004 03:20:01.785655 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39554a25_d378_4002_9359_f07a23f8e621.slice/crio-30a9784f6322c9fd2084db2a45c658bf4b75242cc9574f46173ea1aadd48950e WatchSource:0}: Error finding container 30a9784f6322c9fd2084db2a45c658bf4b75242cc9574f46173ea1aadd48950e: Status 404 returned error can't find the container with id 30a9784f6322c9fd2084db2a45c658bf4b75242cc9574f46173ea1aadd48950e Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.795525 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:20:01 crc kubenswrapper[4770]: I1004 03:20:01.795591 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.481159 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" event={"ID":"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d","Type":"ContainerStarted","Data":"012083d0aaa4b685b2e739797f86d892a0051824113242d4ad28260ee64ff978"} Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.492349 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" event={"ID":"39554a25-d378-4002-9359-f07a23f8e621","Type":"ContainerStarted","Data":"30a9784f6322c9fd2084db2a45c658bf4b75242cc9574f46173ea1aadd48950e"} Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.605172 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-v2tnx"] Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.655280 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-xfb5h"] Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.656418 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.688716 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-xfb5h"] Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.758557 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-config\") pod \"dnsmasq-dns-6c64dcbb65-xfb5h\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.758702 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xscvm\" (UniqueName: \"kubernetes.io/projected/11105f93-4bcd-4fd1-8640-ba20672c5df4-kube-api-access-xscvm\") pod \"dnsmasq-dns-6c64dcbb65-xfb5h\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.758742 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-dns-svc\") pod \"dnsmasq-dns-6c64dcbb65-xfb5h\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.872647 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xscvm\" (UniqueName: \"kubernetes.io/projected/11105f93-4bcd-4fd1-8640-ba20672c5df4-kube-api-access-xscvm\") pod \"dnsmasq-dns-6c64dcbb65-xfb5h\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.873365 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-dns-svc\") pod \"dnsmasq-dns-6c64dcbb65-xfb5h\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.873628 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-config\") pod \"dnsmasq-dns-6c64dcbb65-xfb5h\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.874883 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-dns-svc\") pod \"dnsmasq-dns-6c64dcbb65-xfb5h\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.875385 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-config\") pod \"dnsmasq-dns-6c64dcbb65-xfb5h\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.922583 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xscvm\" (UniqueName: \"kubernetes.io/projected/11105f93-4bcd-4fd1-8640-ba20672c5df4-kube-api-access-xscvm\") pod \"dnsmasq-dns-6c64dcbb65-xfb5h\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:02 crc kubenswrapper[4770]: I1004 03:20:02.997711 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.363475 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-gtdlx"] Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.385114 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-vgk4p"] Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.390668 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.406524 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-vgk4p"] Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.576841 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-xfb5h"] Oct 04 03:20:03 crc kubenswrapper[4770]: W1004 03:20:03.581073 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11105f93_4bcd_4fd1_8640_ba20672c5df4.slice/crio-6043c07f455e1429e74018ba5b598a590bb40046a6bee2322ed80efb00517623 WatchSource:0}: Error finding container 6043c07f455e1429e74018ba5b598a590bb40046a6bee2322ed80efb00517623: Status 404 returned error can't find the container with id 6043c07f455e1429e74018ba5b598a590bb40046a6bee2322ed80efb00517623 Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.590625 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-dns-svc\") pod \"dnsmasq-dns-5d4d9f7875-vgk4p\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.590663 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndc9m\" (UniqueName: \"kubernetes.io/projected/c81c7d43-13ea-4fb9-9f5e-f837e3959603-kube-api-access-ndc9m\") pod \"dnsmasq-dns-5d4d9f7875-vgk4p\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.590695 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-config\") pod \"dnsmasq-dns-5d4d9f7875-vgk4p\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.692171 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-dns-svc\") pod \"dnsmasq-dns-5d4d9f7875-vgk4p\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.692213 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndc9m\" (UniqueName: \"kubernetes.io/projected/c81c7d43-13ea-4fb9-9f5e-f837e3959603-kube-api-access-ndc9m\") pod \"dnsmasq-dns-5d4d9f7875-vgk4p\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.692256 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-config\") pod \"dnsmasq-dns-5d4d9f7875-vgk4p\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.694839 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-dns-svc\") pod \"dnsmasq-dns-5d4d9f7875-vgk4p\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.695110 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-config\") pod \"dnsmasq-dns-5d4d9f7875-vgk4p\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.710564 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndc9m\" (UniqueName: \"kubernetes.io/projected/c81c7d43-13ea-4fb9-9f5e-f837e3959603-kube-api-access-ndc9m\") pod \"dnsmasq-dns-5d4d9f7875-vgk4p\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.731211 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.869992 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.871253 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.873602 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.873813 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.873973 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-qv2nc" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.873974 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.876061 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.876291 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.879173 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.890239 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.997832 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtcmp\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-kube-api-access-xtcmp\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.997899 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.997934 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2def82a8-5927-4ebe-ac87-e8ad232797ee-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.997962 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2def82a8-5927-4ebe-ac87-e8ad232797ee-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.998025 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:03 crc kubenswrapper[4770]: I1004 03:20:03.998056 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:03.998127 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:03.998147 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:03.998181 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:03.998214 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:03.998271 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099099 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099190 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtcmp\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-kube-api-access-xtcmp\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099223 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099248 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2def82a8-5927-4ebe-ac87-e8ad232797ee-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099272 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2def82a8-5927-4ebe-ac87-e8ad232797ee-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099309 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099328 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099357 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099371 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099394 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099416 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.099909 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.100118 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.100403 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.100426 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.100737 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.105229 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.105715 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.108940 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2def82a8-5927-4ebe-ac87-e8ad232797ee-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.109045 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.115936 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2def82a8-5927-4ebe-ac87-e8ad232797ee-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.117949 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtcmp\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-kube-api-access-xtcmp\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.126694 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.201246 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.245868 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-vgk4p"] Oct 04 03:20:04 crc kubenswrapper[4770]: W1004 03:20:04.259178 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc81c7d43_13ea_4fb9_9f5e_f837e3959603.slice/crio-d271338295de3aa6f508507bec5f4bf458c936bb901478f304dbe351fe3fc9a7 WatchSource:0}: Error finding container d271338295de3aa6f508507bec5f4bf458c936bb901478f304dbe351fe3fc9a7: Status 404 returned error can't find the container with id d271338295de3aa6f508507bec5f4bf458c936bb901478f304dbe351fe3fc9a7 Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.511401 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" event={"ID":"11105f93-4bcd-4fd1-8640-ba20672c5df4","Type":"ContainerStarted","Data":"6043c07f455e1429e74018ba5b598a590bb40046a6bee2322ed80efb00517623"} Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.513611 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" event={"ID":"c81c7d43-13ea-4fb9-9f5e-f837e3959603","Type":"ContainerStarted","Data":"d271338295de3aa6f508507bec5f4bf458c936bb901478f304dbe351fe3fc9a7"} Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.537835 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.544861 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.547696 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-l9q7d" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.547850 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.548028 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.548138 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.548243 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.548374 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.548573 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.559272 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.711690 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712140 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/642fe06d-e5b8-461e-a1f7-dbcefc071945-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712170 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/642fe06d-e5b8-461e-a1f7-dbcefc071945-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712189 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gcjw\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-kube-api-access-4gcjw\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712244 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712272 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712295 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712317 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712341 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712361 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.712387 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.763708 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813662 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/642fe06d-e5b8-461e-a1f7-dbcefc071945-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813712 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813732 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/642fe06d-e5b8-461e-a1f7-dbcefc071945-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813785 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gcjw\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-kube-api-access-4gcjw\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813836 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813854 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813875 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813894 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813928 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813948 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.813971 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.814985 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.815054 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.815626 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.815802 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.823137 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.824679 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/642fe06d-e5b8-461e-a1f7-dbcefc071945-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.825096 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.827975 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/642fe06d-e5b8-461e-a1f7-dbcefc071945-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.828097 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: W1004 03:20:04.830237 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2def82a8_5927_4ebe_ac87_e8ad232797ee.slice/crio-b7998530f1dfa56d5f6be2c686686a50395bb8d934c54b5975ea3d450814dcf5 WatchSource:0}: Error finding container b7998530f1dfa56d5f6be2c686686a50395bb8d934c54b5975ea3d450814dcf5: Status 404 returned error can't find the container with id b7998530f1dfa56d5f6be2c686686a50395bb8d934c54b5975ea3d450814dcf5 Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.830973 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gcjw\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-kube-api-access-4gcjw\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.841550 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.843621 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:04 crc kubenswrapper[4770]: I1004 03:20:04.870881 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:05 crc kubenswrapper[4770]: I1004 03:20:05.517435 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:20:05 crc kubenswrapper[4770]: W1004 03:20:05.529025 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod642fe06d_e5b8_461e_a1f7_dbcefc071945.slice/crio-e3aab1fd6e666025903602376b35fc475759dabf57ddeba8d7c0b7252952aa50 WatchSource:0}: Error finding container e3aab1fd6e666025903602376b35fc475759dabf57ddeba8d7c0b7252952aa50: Status 404 returned error can't find the container with id e3aab1fd6e666025903602376b35fc475759dabf57ddeba8d7c0b7252952aa50 Oct 04 03:20:05 crc kubenswrapper[4770]: I1004 03:20:05.536093 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2def82a8-5927-4ebe-ac87-e8ad232797ee","Type":"ContainerStarted","Data":"b7998530f1dfa56d5f6be2c686686a50395bb8d934c54b5975ea3d450814dcf5"} Oct 04 03:20:06 crc kubenswrapper[4770]: I1004 03:20:06.572054 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"642fe06d-e5b8-461e-a1f7-dbcefc071945","Type":"ContainerStarted","Data":"e3aab1fd6e666025903602376b35fc475759dabf57ddeba8d7c0b7252952aa50"} Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.411037 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.415066 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.424435 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.428471 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.428528 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xjntg" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.428893 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.429617 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.431776 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.431451 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.545478 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.548878 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.552084 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-9h5sm" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.552163 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.552444 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.552531 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.559813 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.559864 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kolla-config\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.559917 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.559960 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-default\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.559993 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtg8h\" (UniqueName: \"kubernetes.io/projected/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kube-api-access-xtg8h\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.560066 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.560086 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.560102 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-secrets\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.560140 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.570814 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661344 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kolla-config\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661406 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661483 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661501 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661524 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661550 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-default\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661569 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gktjh\" (UniqueName: \"kubernetes.io/projected/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kube-api-access-gktjh\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661595 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661614 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661633 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtg8h\" (UniqueName: \"kubernetes.io/projected/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kube-api-access-xtg8h\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661667 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661684 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661701 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661718 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661733 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-secrets\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661752 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.661804 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.662600 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kolla-config\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.663437 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.663686 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.664050 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-default\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.666053 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.679778 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-secrets\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.680134 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.693167 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.715753 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtg8h\" (UniqueName: \"kubernetes.io/projected/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kube-api-access-xtg8h\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.729201 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.741866 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.763427 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.763477 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.763509 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gktjh\" (UniqueName: \"kubernetes.io/projected/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kube-api-access-gktjh\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.763536 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.763553 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.763591 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.763607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.763640 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.763676 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.764285 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.765120 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.766366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.767507 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.768156 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.768777 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.772463 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.772594 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.792408 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gktjh\" (UniqueName: \"kubernetes.io/projected/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kube-api-access-gktjh\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.848400 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.867992 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.869725 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.873965 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.882288 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-vvpdw" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.882494 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.882898 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.887034 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.968063 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq24q\" (UniqueName: \"kubernetes.io/projected/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kube-api-access-rq24q\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.968137 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.968187 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kolla-config\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.968208 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:07 crc kubenswrapper[4770]: I1004 03:20:07.968552 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-config-data\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.070742 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq24q\" (UniqueName: \"kubernetes.io/projected/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kube-api-access-rq24q\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.070822 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.070862 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kolla-config\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.070889 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.070911 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-config-data\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.072167 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-config-data\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.072242 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kolla-config\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.079788 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.089065 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.116950 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq24q\" (UniqueName: \"kubernetes.io/projected/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kube-api-access-rq24q\") pod \"memcached-0\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " pod="openstack/memcached-0" Oct 04 03:20:08 crc kubenswrapper[4770]: I1004 03:20:08.220401 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 03:20:09 crc kubenswrapper[4770]: I1004 03:20:09.324955 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:20:09 crc kubenswrapper[4770]: I1004 03:20:09.327125 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:20:09 crc kubenswrapper[4770]: I1004 03:20:09.329524 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:20:09 crc kubenswrapper[4770]: I1004 03:20:09.332403 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-r9d55" Oct 04 03:20:09 crc kubenswrapper[4770]: I1004 03:20:09.402990 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpm2f\" (UniqueName: \"kubernetes.io/projected/60a374b7-7611-4f40-b0ae-a6238903e0ea-kube-api-access-vpm2f\") pod \"kube-state-metrics-0\" (UID: \"60a374b7-7611-4f40-b0ae-a6238903e0ea\") " pod="openstack/kube-state-metrics-0" Oct 04 03:20:09 crc kubenswrapper[4770]: I1004 03:20:09.504213 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpm2f\" (UniqueName: \"kubernetes.io/projected/60a374b7-7611-4f40-b0ae-a6238903e0ea-kube-api-access-vpm2f\") pod \"kube-state-metrics-0\" (UID: \"60a374b7-7611-4f40-b0ae-a6238903e0ea\") " pod="openstack/kube-state-metrics-0" Oct 04 03:20:09 crc kubenswrapper[4770]: I1004 03:20:09.523099 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpm2f\" (UniqueName: \"kubernetes.io/projected/60a374b7-7611-4f40-b0ae-a6238903e0ea-kube-api-access-vpm2f\") pod \"kube-state-metrics-0\" (UID: \"60a374b7-7611-4f40-b0ae-a6238903e0ea\") " pod="openstack/kube-state-metrics-0" Oct 04 03:20:09 crc kubenswrapper[4770]: I1004 03:20:09.652941 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-r9d55" Oct 04 03:20:09 crc kubenswrapper[4770]: I1004 03:20:09.660405 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.234070 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-c2v9x"] Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.235857 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.238957 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-vr4gz" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.239415 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.239499 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.243312 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6nz6q"] Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.245816 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.252756 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-c2v9x"] Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.271794 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6nz6q"] Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.374128 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-run\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.374374 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-lib\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.374475 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-log-ovn\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.374585 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.374655 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-etc-ovs\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.374742 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run-ovn\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.374819 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5f2e918-133c-440e-ab52-2a389ddc9ede-scripts\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.374904 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-ovn-controller-tls-certs\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.375088 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-combined-ca-bundle\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.375169 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glpwm\" (UniqueName: \"kubernetes.io/projected/c5f2e918-133c-440e-ab52-2a389ddc9ede-kube-api-access-glpwm\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.375237 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58bkz\" (UniqueName: \"kubernetes.io/projected/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-kube-api-access-58bkz\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.375326 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-scripts\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.375369 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-log\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476497 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5f2e918-133c-440e-ab52-2a389ddc9ede-scripts\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476546 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-ovn-controller-tls-certs\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476577 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-combined-ca-bundle\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glpwm\" (UniqueName: \"kubernetes.io/projected/c5f2e918-133c-440e-ab52-2a389ddc9ede-kube-api-access-glpwm\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476633 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58bkz\" (UniqueName: \"kubernetes.io/projected/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-kube-api-access-58bkz\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476674 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-scripts\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476691 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-log\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476730 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-run\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476751 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-lib\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476777 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-log-ovn\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476795 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476810 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-etc-ovs\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.476825 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run-ovn\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.477376 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run-ovn\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.477639 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-log\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.477673 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-etc-ovs\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.477687 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-run\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.477681 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.477732 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-log-ovn\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.477938 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-lib\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.479590 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5f2e918-133c-440e-ab52-2a389ddc9ede-scripts\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.481726 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-scripts\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.492707 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-combined-ca-bundle\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.503856 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-ovn-controller-tls-certs\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.518637 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glpwm\" (UniqueName: \"kubernetes.io/projected/c5f2e918-133c-440e-ab52-2a389ddc9ede-kube-api-access-glpwm\") pod \"ovn-controller-ovs-6nz6q\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.525760 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58bkz\" (UniqueName: \"kubernetes.io/projected/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-kube-api-access-58bkz\") pod \"ovn-controller-c2v9x\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.553369 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:13 crc kubenswrapper[4770]: I1004 03:20:13.561342 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.120485 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.122702 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.126984 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.127448 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.127500 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.127740 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.128123 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-6jcf6" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.130915 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.194114 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.194172 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.194224 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-config\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.194249 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.194273 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.195286 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtpmm\" (UniqueName: \"kubernetes.io/projected/57f70627-2bdd-4780-8044-75c3de4aee05-kube-api-access-jtpmm\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.198159 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.198334 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.299861 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtpmm\" (UniqueName: \"kubernetes.io/projected/57f70627-2bdd-4780-8044-75c3de4aee05-kube-api-access-jtpmm\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.299927 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.299983 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.300040 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.300070 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.300109 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-config\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.300134 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.300161 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.303603 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.304162 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-config\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.304641 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.309333 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.311273 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.316047 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.321537 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtpmm\" (UniqueName: \"kubernetes.io/projected/57f70627-2bdd-4780-8044-75c3de4aee05-kube-api-access-jtpmm\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.324819 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.340675 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:14 crc kubenswrapper[4770]: I1004 03:20:14.458401 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.331203 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.335073 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.340603 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.340991 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.342183 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vpmpc" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.342445 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.349230 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.454950 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.455065 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.455123 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-config\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.455162 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxng7\" (UniqueName: \"kubernetes.io/projected/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-kube-api-access-rxng7\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.455229 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.455267 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.455342 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.455385 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.557038 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.557130 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-config\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.557157 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxng7\" (UniqueName: \"kubernetes.io/projected/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-kube-api-access-rxng7\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.557996 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.559285 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-config\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.559433 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.559080 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.568860 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.568930 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.568991 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.569367 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.569836 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.575615 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.576439 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxng7\" (UniqueName: \"kubernetes.io/projected/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-kube-api-access-rxng7\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.577421 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.582760 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.584582 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:17 crc kubenswrapper[4770]: I1004 03:20:17.678708 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.694363 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.694997 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ndc9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5d4d9f7875-vgk4p_openstack(c81c7d43-13ea-4fb9-9f5e-f837e3959603): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.696281 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" podUID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.723281 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.723451 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xscvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6c64dcbb65-xfb5h_openstack(11105f93-4bcd-4fd1-8640-ba20672c5df4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.724793 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.724886 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cp54d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-546d69f86c-v2tnx_openstack(dbd533f1-d82e-4ecf-9a6d-920d70d52d7d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.724939 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" podUID="11105f93-4bcd-4fd1-8640-ba20672c5df4" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.726043 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" podUID="dbd533f1-d82e-4ecf-9a6d-920d70d52d7d" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.756871 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.757206 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7lkg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7f9579fb85-gtdlx_openstack(39554a25-d378-4002-9359-f07a23f8e621): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.762432 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" podUID="39554a25-d378-4002-9359-f07a23f8e621" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.774983 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26\\\"\"" pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" podUID="11105f93-4bcd-4fd1-8640-ba20672c5df4" Oct 04 03:20:22 crc kubenswrapper[4770]: E1004 03:20:22.775256 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26\\\"\"" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" podUID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.271730 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.299457 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.386041 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp54d\" (UniqueName: \"kubernetes.io/projected/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-kube-api-access-cp54d\") pod \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\" (UID: \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\") " Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.386382 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-config\") pod \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\" (UID: \"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d\") " Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.386932 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-config" (OuterVolumeSpecName: "config") pod "dbd533f1-d82e-4ecf-9a6d-920d70d52d7d" (UID: "dbd533f1-d82e-4ecf-9a6d-920d70d52d7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.394040 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-kube-api-access-cp54d" (OuterVolumeSpecName: "kube-api-access-cp54d") pod "dbd533f1-d82e-4ecf-9a6d-920d70d52d7d" (UID: "dbd533f1-d82e-4ecf-9a6d-920d70d52d7d"). InnerVolumeSpecName "kube-api-access-cp54d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.488201 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp54d\" (UniqueName: \"kubernetes.io/projected/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-kube-api-access-cp54d\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.488244 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.607042 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.632376 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.650405 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 03:20:23 crc kubenswrapper[4770]: W1004 03:20:23.663032 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60a374b7_7611_4f40_b0ae_a6238903e0ea.slice/crio-5b816be7aa73f09f3b7105f982c4bde6cfd4b19cff32c9048a336134b52a5696 WatchSource:0}: Error finding container 5b816be7aa73f09f3b7105f982c4bde6cfd4b19cff32c9048a336134b52a5696: Status 404 returned error can't find the container with id 5b816be7aa73f09f3b7105f982c4bde6cfd4b19cff32c9048a336134b52a5696 Oct 04 03:20:23 crc kubenswrapper[4770]: W1004 03:20:23.664145 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41309df4_53d7_42a8_8a5b_4816cc25fdfb.slice/crio-e43aa5b60dcc68b523a2831b338d417300dba571425330b9d39c087939b534fe WatchSource:0}: Error finding container e43aa5b60dcc68b523a2831b338d417300dba571425330b9d39c087939b534fe: Status 404 returned error can't find the container with id e43aa5b60dcc68b523a2831b338d417300dba571425330b9d39c087939b534fe Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.666055 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-c2v9x"] Oct 04 03:20:23 crc kubenswrapper[4770]: W1004 03:20:23.667255 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35f43e87_e3d9_4508_8bb1_4fe89c64017f.slice/crio-3f95e0d8695bd0efa45dc9005e6070c12f2bd8e488946031bf4b233e2927349d WatchSource:0}: Error finding container 3f95e0d8695bd0efa45dc9005e6070c12f2bd8e488946031bf4b233e2927349d: Status 404 returned error can't find the container with id 3f95e0d8695bd0efa45dc9005e6070c12f2bd8e488946031bf4b233e2927349d Oct 04 03:20:23 crc kubenswrapper[4770]: W1004 03:20:23.668644 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7916b1d6_76ad_4780_bfc2_2a0a738b8b8d.slice/crio-6d25180e1bdf6781040961e6aad0cc608329ce28985693de6216fcd4475a20cc WatchSource:0}: Error finding container 6d25180e1bdf6781040961e6aad0cc608329ce28985693de6216fcd4475a20cc: Status 404 returned error can't find the container with id 6d25180e1bdf6781040961e6aad0cc608329ce28985693de6216fcd4475a20cc Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.779297 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"41309df4-53d7-42a8-8a5b-4816cc25fdfb","Type":"ContainerStarted","Data":"e43aa5b60dcc68b523a2831b338d417300dba571425330b9d39c087939b534fe"} Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.781375 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7","Type":"ContainerStarted","Data":"72555b469a440d5f31b527b5d35fb623f29918672fdfd1486433293d3dbd19c1"} Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.782628 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"35f43e87-e3d9-4508-8bb1-4fe89c64017f","Type":"ContainerStarted","Data":"3f95e0d8695bd0efa45dc9005e6070c12f2bd8e488946031bf4b233e2927349d"} Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.783595 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" event={"ID":"dbd533f1-d82e-4ecf-9a6d-920d70d52d7d","Type":"ContainerDied","Data":"012083d0aaa4b685b2e739797f86d892a0051824113242d4ad28260ee64ff978"} Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.783663 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-v2tnx" Oct 04 03:20:23 crc kubenswrapper[4770]: W1004 03:20:23.785932 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57f70627_2bdd_4780_8044_75c3de4aee05.slice/crio-1cb22383ea5a902e5e9d36f42af1e33ae0907496e90bda94cd7c9317731febe7 WatchSource:0}: Error finding container 1cb22383ea5a902e5e9d36f42af1e33ae0907496e90bda94cd7c9317731febe7: Status 404 returned error can't find the container with id 1cb22383ea5a902e5e9d36f42af1e33ae0907496e90bda94cd7c9317731febe7 Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.786182 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x" event={"ID":"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d","Type":"ContainerStarted","Data":"6d25180e1bdf6781040961e6aad0cc608329ce28985693de6216fcd4475a20cc"} Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.790414 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.791916 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"60a374b7-7611-4f40-b0ae-a6238903e0ea","Type":"ContainerStarted","Data":"5b816be7aa73f09f3b7105f982c4bde6cfd4b19cff32c9048a336134b52a5696"} Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.866728 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-v2tnx"] Oct 04 03:20:23 crc kubenswrapper[4770]: I1004 03:20:23.872330 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-v2tnx"] Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.119209 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.207550 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-dns-svc\") pod \"39554a25-d378-4002-9359-f07a23f8e621\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.207680 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-config\") pod \"39554a25-d378-4002-9359-f07a23f8e621\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.207758 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7lkg\" (UniqueName: \"kubernetes.io/projected/39554a25-d378-4002-9359-f07a23f8e621-kube-api-access-t7lkg\") pod \"39554a25-d378-4002-9359-f07a23f8e621\" (UID: \"39554a25-d378-4002-9359-f07a23f8e621\") " Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.208326 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-config" (OuterVolumeSpecName: "config") pod "39554a25-d378-4002-9359-f07a23f8e621" (UID: "39554a25-d378-4002-9359-f07a23f8e621"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.208754 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39554a25-d378-4002-9359-f07a23f8e621" (UID: "39554a25-d378-4002-9359-f07a23f8e621"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.219236 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39554a25-d378-4002-9359-f07a23f8e621-kube-api-access-t7lkg" (OuterVolumeSpecName: "kube-api-access-t7lkg") pod "39554a25-d378-4002-9359-f07a23f8e621" (UID: "39554a25-d378-4002-9359-f07a23f8e621"). InnerVolumeSpecName "kube-api-access-t7lkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.309480 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.309530 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7lkg\" (UniqueName: \"kubernetes.io/projected/39554a25-d378-4002-9359-f07a23f8e621-kube-api-access-t7lkg\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.309549 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39554a25-d378-4002-9359-f07a23f8e621-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.464624 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 03:20:24 crc kubenswrapper[4770]: W1004 03:20:24.502890 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4508feb_4ff8_47ec_86d7_7a9a0ad61a24.slice/crio-287d87a84cf1bd29f97b9071e57d29a0be857b26458c07be1832624e83dcc69b WatchSource:0}: Error finding container 287d87a84cf1bd29f97b9071e57d29a0be857b26458c07be1832624e83dcc69b: Status 404 returned error can't find the container with id 287d87a84cf1bd29f97b9071e57d29a0be857b26458c07be1832624e83dcc69b Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.545613 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6nz6q"] Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.802156 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2def82a8-5927-4ebe-ac87-e8ad232797ee","Type":"ContainerStarted","Data":"5305bb132085306589edf135617a3aa23523ea917b7840da309c820a5a534359"} Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.809703 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" event={"ID":"39554a25-d378-4002-9359-f07a23f8e621","Type":"ContainerDied","Data":"30a9784f6322c9fd2084db2a45c658bf4b75242cc9574f46173ea1aadd48950e"} Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.809704 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-gtdlx" Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.811749 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24","Type":"ContainerStarted","Data":"287d87a84cf1bd29f97b9071e57d29a0be857b26458c07be1832624e83dcc69b"} Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.818993 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"642fe06d-e5b8-461e-a1f7-dbcefc071945","Type":"ContainerStarted","Data":"0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01"} Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.820809 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"57f70627-2bdd-4780-8044-75c3de4aee05","Type":"ContainerStarted","Data":"1cb22383ea5a902e5e9d36f42af1e33ae0907496e90bda94cd7c9317731febe7"} Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.909388 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-gtdlx"] Oct 04 03:20:24 crc kubenswrapper[4770]: I1004 03:20:24.914085 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-gtdlx"] Oct 04 03:20:24 crc kubenswrapper[4770]: W1004 03:20:24.946825 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5f2e918_133c_440e_ab52_2a389ddc9ede.slice/crio-7dcea88773c57dcac7ec8f2a5d3895d9156de475966a38c28b29c007a198d6db WatchSource:0}: Error finding container 7dcea88773c57dcac7ec8f2a5d3895d9156de475966a38c28b29c007a198d6db: Status 404 returned error can't find the container with id 7dcea88773c57dcac7ec8f2a5d3895d9156de475966a38c28b29c007a198d6db Oct 04 03:20:25 crc kubenswrapper[4770]: I1004 03:20:25.683242 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39554a25-d378-4002-9359-f07a23f8e621" path="/var/lib/kubelet/pods/39554a25-d378-4002-9359-f07a23f8e621/volumes" Oct 04 03:20:25 crc kubenswrapper[4770]: I1004 03:20:25.684364 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbd533f1-d82e-4ecf-9a6d-920d70d52d7d" path="/var/lib/kubelet/pods/dbd533f1-d82e-4ecf-9a6d-920d70d52d7d/volumes" Oct 04 03:20:25 crc kubenswrapper[4770]: I1004 03:20:25.869433 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6nz6q" event={"ID":"c5f2e918-133c-440e-ab52-2a389ddc9ede","Type":"ContainerStarted","Data":"7dcea88773c57dcac7ec8f2a5d3895d9156de475966a38c28b29c007a198d6db"} Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.795407 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.796617 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.937482 4770 generic.go:334] "Generic (PLEG): container finished" podID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerID="b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4" exitCode=0 Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.937563 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6nz6q" event={"ID":"c5f2e918-133c-440e-ab52-2a389ddc9ede","Type":"ContainerDied","Data":"b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4"} Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.943557 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7","Type":"ContainerStarted","Data":"6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d"} Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.947128 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"35f43e87-e3d9-4508-8bb1-4fe89c64017f","Type":"ContainerStarted","Data":"cfa890ec4ee2fe7f12e80ec7872bbb31140af8442fe957e965fd9230cc298d8d"} Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.957653 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"57f70627-2bdd-4780-8044-75c3de4aee05","Type":"ContainerStarted","Data":"92885b43eda03ead2301755a74384e53d3b27a5aa4fddd4305ff2576a9f59a7c"} Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.961636 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x" event={"ID":"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d","Type":"ContainerStarted","Data":"37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30"} Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.961878 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-c2v9x" Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.970131 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"60a374b7-7611-4f40-b0ae-a6238903e0ea","Type":"ContainerStarted","Data":"6633e392a5a6f2b64dfbad7667d1b0267f86fde4b17e7db1b8808041de34c634"} Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.970819 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.972977 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24","Type":"ContainerStarted","Data":"ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a"} Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.974799 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"41309df4-53d7-42a8-8a5b-4816cc25fdfb","Type":"ContainerStarted","Data":"5005796694c28178137720d1fa9150a9692d20f97c8d47bfb7d90339bebdf44b"} Oct 04 03:20:31 crc kubenswrapper[4770]: I1004 03:20:31.975216 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 04 03:20:32 crc kubenswrapper[4770]: I1004 03:20:32.042319 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-c2v9x" podStartSLOduration=11.827602449 podStartE2EDuration="19.042289507s" podCreationTimestamp="2025-10-04 03:20:13 +0000 UTC" firstStartedPulling="2025-10-04 03:20:23.672534716 +0000 UTC m=+1034.964544428" lastFinishedPulling="2025-10-04 03:20:30.887221774 +0000 UTC m=+1042.179231486" observedRunningTime="2025-10-04 03:20:32.036171117 +0000 UTC m=+1043.328180829" watchObservedRunningTime="2025-10-04 03:20:32.042289507 +0000 UTC m=+1043.334299229" Oct 04 03:20:32 crc kubenswrapper[4770]: I1004 03:20:32.059383 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.756850277 podStartE2EDuration="23.059363154s" podCreationTimestamp="2025-10-04 03:20:09 +0000 UTC" firstStartedPulling="2025-10-04 03:20:23.665801199 +0000 UTC m=+1034.957810921" lastFinishedPulling="2025-10-04 03:20:30.968314076 +0000 UTC m=+1042.260323798" observedRunningTime="2025-10-04 03:20:32.053651934 +0000 UTC m=+1043.345661686" watchObservedRunningTime="2025-10-04 03:20:32.059363154 +0000 UTC m=+1043.351372876" Oct 04 03:20:32 crc kubenswrapper[4770]: I1004 03:20:32.078895 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=18.330005418 podStartE2EDuration="25.078878955s" podCreationTimestamp="2025-10-04 03:20:07 +0000 UTC" firstStartedPulling="2025-10-04 03:20:23.665879011 +0000 UTC m=+1034.957888723" lastFinishedPulling="2025-10-04 03:20:30.414752548 +0000 UTC m=+1041.706762260" observedRunningTime="2025-10-04 03:20:32.077627822 +0000 UTC m=+1043.369637544" watchObservedRunningTime="2025-10-04 03:20:32.078878955 +0000 UTC m=+1043.370888667" Oct 04 03:20:32 crc kubenswrapper[4770]: I1004 03:20:32.986513 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6nz6q" event={"ID":"c5f2e918-133c-440e-ab52-2a389ddc9ede","Type":"ContainerStarted","Data":"51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c"} Oct 04 03:20:32 crc kubenswrapper[4770]: I1004 03:20:32.986952 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6nz6q" event={"ID":"c5f2e918-133c-440e-ab52-2a389ddc9ede","Type":"ContainerStarted","Data":"4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd"} Oct 04 03:20:33 crc kubenswrapper[4770]: I1004 03:20:33.017423 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6nz6q" podStartSLOduration=14.086174356 podStartE2EDuration="20.01738591s" podCreationTimestamp="2025-10-04 03:20:13 +0000 UTC" firstStartedPulling="2025-10-04 03:20:24.950130896 +0000 UTC m=+1036.242140608" lastFinishedPulling="2025-10-04 03:20:30.88134246 +0000 UTC m=+1042.173352162" observedRunningTime="2025-10-04 03:20:33.008695002 +0000 UTC m=+1044.300704714" watchObservedRunningTime="2025-10-04 03:20:33.01738591 +0000 UTC m=+1044.309395622" Oct 04 03:20:33 crc kubenswrapper[4770]: I1004 03:20:33.562420 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:33 crc kubenswrapper[4770]: I1004 03:20:33.562471 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.010310 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24","Type":"ContainerStarted","Data":"347a7235cf642c5bee0dc831171a48e6731de85a58d10232d67b2b94de43b821"} Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.012840 4770 generic.go:334] "Generic (PLEG): container finished" podID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" containerID="6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d" exitCode=0 Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.012920 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7","Type":"ContainerDied","Data":"6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d"} Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.014631 4770 generic.go:334] "Generic (PLEG): container finished" podID="35f43e87-e3d9-4508-8bb1-4fe89c64017f" containerID="cfa890ec4ee2fe7f12e80ec7872bbb31140af8442fe957e965fd9230cc298d8d" exitCode=0 Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.014700 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"35f43e87-e3d9-4508-8bb1-4fe89c64017f","Type":"ContainerDied","Data":"cfa890ec4ee2fe7f12e80ec7872bbb31140af8442fe957e965fd9230cc298d8d"} Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.017295 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"57f70627-2bdd-4780-8044-75c3de4aee05","Type":"ContainerStarted","Data":"760e8cc98a032092f27104dafba09f778ecc47ca79026332349329b345264e45"} Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.042258 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=9.066866062 podStartE2EDuration="19.042232928s" podCreationTimestamp="2025-10-04 03:20:16 +0000 UTC" firstStartedPulling="2025-10-04 03:20:24.507044019 +0000 UTC m=+1035.799053751" lastFinishedPulling="2025-10-04 03:20:34.482410905 +0000 UTC m=+1045.774420617" observedRunningTime="2025-10-04 03:20:35.038306895 +0000 UTC m=+1046.330316607" watchObservedRunningTime="2025-10-04 03:20:35.042232928 +0000 UTC m=+1046.334242630" Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.123657 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.43109025 podStartE2EDuration="22.123631338s" podCreationTimestamp="2025-10-04 03:20:13 +0000 UTC" firstStartedPulling="2025-10-04 03:20:23.790126114 +0000 UTC m=+1035.082135836" lastFinishedPulling="2025-10-04 03:20:34.482667202 +0000 UTC m=+1045.774676924" observedRunningTime="2025-10-04 03:20:35.108288377 +0000 UTC m=+1046.400298089" watchObservedRunningTime="2025-10-04 03:20:35.123631338 +0000 UTC m=+1046.415641050" Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.459637 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.501786 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.700740 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:35 crc kubenswrapper[4770]: I1004 03:20:35.734044 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.029451 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7","Type":"ContainerStarted","Data":"6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3"} Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.033313 4770 generic.go:334] "Generic (PLEG): container finished" podID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" containerID="e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5" exitCode=0 Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.033444 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" event={"ID":"c81c7d43-13ea-4fb9-9f5e-f837e3959603","Type":"ContainerDied","Data":"e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5"} Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.036260 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"35f43e87-e3d9-4508-8bb1-4fe89c64017f","Type":"ContainerStarted","Data":"0218e06d6c52fad79e58c3f1c9b92b86ec1c9c65d56610da75164f0326b481a0"} Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.036966 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.037088 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.076740 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.837486353 podStartE2EDuration="30.076694494s" podCreationTimestamp="2025-10-04 03:20:06 +0000 UTC" firstStartedPulling="2025-10-04 03:20:23.299952753 +0000 UTC m=+1034.591962455" lastFinishedPulling="2025-10-04 03:20:30.539160884 +0000 UTC m=+1041.831170596" observedRunningTime="2025-10-04 03:20:36.070065281 +0000 UTC m=+1047.362074993" watchObservedRunningTime="2025-10-04 03:20:36.076694494 +0000 UTC m=+1047.368704206" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.119356 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.122148 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.135090 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=22.926256056 podStartE2EDuration="30.135065651s" podCreationTimestamp="2025-10-04 03:20:06 +0000 UTC" firstStartedPulling="2025-10-04 03:20:23.669509766 +0000 UTC m=+1034.961519478" lastFinishedPulling="2025-10-04 03:20:30.878319361 +0000 UTC m=+1042.170329073" observedRunningTime="2025-10-04 03:20:36.123069748 +0000 UTC m=+1047.415079490" watchObservedRunningTime="2025-10-04 03:20:36.135065651 +0000 UTC m=+1047.427075373" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.455686 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-xfb5h"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.513909 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-zjhc4"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.515886 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.519445 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-tn765"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.521362 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.522653 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.524541 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.540899 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-zjhc4"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.550312 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-tn765"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675458 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675517 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-ovsdbserver-nb\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675539 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-dns-svc\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675582 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-config\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675600 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfg9\" (UniqueName: \"kubernetes.io/projected/4b14dd83-328d-45dc-baa0-b63afe4e03a5-kube-api-access-8dfg9\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675632 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4444038-6604-4e51-b27d-20d3616bf640-config\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675656 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovs-rundir\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675673 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovn-rundir\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675688 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-combined-ca-bundle\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.675728 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh8bv\" (UniqueName: \"kubernetes.io/projected/b4444038-6604-4e51-b27d-20d3616bf640-kube-api-access-kh8bv\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.703163 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-vgk4p"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.758500 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.759686 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.773906 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.774139 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.774276 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-tqjrq" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.774387 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.777793 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-config\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.777850 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfg9\" (UniqueName: \"kubernetes.io/projected/4b14dd83-328d-45dc-baa0-b63afe4e03a5-kube-api-access-8dfg9\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.777902 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4444038-6604-4e51-b27d-20d3616bf640-config\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.777944 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovs-rundir\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.777977 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovn-rundir\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.777994 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-combined-ca-bundle\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.778087 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh8bv\" (UniqueName: \"kubernetes.io/projected/b4444038-6604-4e51-b27d-20d3616bf640-kube-api-access-kh8bv\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.778131 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.778246 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-ovsdbserver-nb\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.778275 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-dns-svc\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.779386 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-dns-svc\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.779956 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-config\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.781694 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4444038-6604-4e51-b27d-20d3616bf640-config\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.782069 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovs-rundir\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.782136 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovn-rundir\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.783902 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-ovsdbserver-nb\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.795243 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-lbkqb"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.797026 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.798337 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-combined-ca-bundle\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.807073 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.814367 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfg9\" (UniqueName: \"kubernetes.io/projected/4b14dd83-328d-45dc-baa0-b63afe4e03a5-kube-api-access-8dfg9\") pod \"dnsmasq-dns-6b8749979c-zjhc4\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.814773 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.815362 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.819028 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh8bv\" (UniqueName: \"kubernetes.io/projected/b4444038-6604-4e51-b27d-20d3616bf640-kube-api-access-kh8bv\") pod \"ovn-controller-metrics-tn765\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.829069 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-lbkqb"] Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.859831 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.876681 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889180 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889249 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-nb\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889307 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-config\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889343 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-sb\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889457 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-scripts\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889494 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889514 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889564 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4f29\" (UniqueName: \"kubernetes.io/projected/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-kube-api-access-w4f29\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889596 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-config\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889638 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889690 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph4mp\" (UniqueName: \"kubernetes.io/projected/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-kube-api-access-ph4mp\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.889713 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-dns-svc\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.979847 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.998789 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-config\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.998847 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-sb\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.998885 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-scripts\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.998905 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.998920 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.998959 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4f29\" (UniqueName: \"kubernetes.io/projected/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-kube-api-access-w4f29\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:36 crc kubenswrapper[4770]: I1004 03:20:36.998981 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-config\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:36.998998 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.006964 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-scripts\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.007123 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-sb\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.007164 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-config\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.007251 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph4mp\" (UniqueName: \"kubernetes.io/projected/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-kube-api-access-ph4mp\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.007284 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-dns-svc\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.007339 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.007381 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-nb\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.011114 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-config\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.013954 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.016300 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-dns-svc\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.018992 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-nb\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.021120 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.026624 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.031427 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph4mp\" (UniqueName: \"kubernetes.io/projected/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-kube-api-access-ph4mp\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.035584 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4f29\" (UniqueName: \"kubernetes.io/projected/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-kube-api-access-w4f29\") pod \"dnsmasq-dns-8454ffc489-lbkqb\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.037546 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.075724 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" event={"ID":"11105f93-4bcd-4fd1-8640-ba20672c5df4","Type":"ContainerDied","Data":"6043c07f455e1429e74018ba5b598a590bb40046a6bee2322ed80efb00517623"} Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.075960 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c64dcbb65-xfb5h" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.115509 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-config\") pod \"11105f93-4bcd-4fd1-8640-ba20672c5df4\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.115693 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-dns-svc\") pod \"11105f93-4bcd-4fd1-8640-ba20672c5df4\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.115908 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xscvm\" (UniqueName: \"kubernetes.io/projected/11105f93-4bcd-4fd1-8640-ba20672c5df4-kube-api-access-xscvm\") pod \"11105f93-4bcd-4fd1-8640-ba20672c5df4\" (UID: \"11105f93-4bcd-4fd1-8640-ba20672c5df4\") " Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.118836 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11105f93-4bcd-4fd1-8640-ba20672c5df4" (UID: "11105f93-4bcd-4fd1-8640-ba20672c5df4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.123216 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11105f93-4bcd-4fd1-8640-ba20672c5df4-kube-api-access-xscvm" (OuterVolumeSpecName: "kube-api-access-xscvm") pod "11105f93-4bcd-4fd1-8640-ba20672c5df4" (UID: "11105f93-4bcd-4fd1-8640-ba20672c5df4"). InnerVolumeSpecName "kube-api-access-xscvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.219106 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.219135 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xscvm\" (UniqueName: \"kubernetes.io/projected/11105f93-4bcd-4fd1-8640-ba20672c5df4-kube-api-access-xscvm\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.250176 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.266099 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.341791 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-zjhc4"] Oct 04 03:20:37 crc kubenswrapper[4770]: W1004 03:20:37.354559 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b14dd83_328d_45dc_baa0_b63afe4e03a5.slice/crio-1fcb384cb07b3486eafaabe6e7406da25f1e2032a6dfc2eba058a1fcc74bb8ef WatchSource:0}: Error finding container 1fcb384cb07b3486eafaabe6e7406da25f1e2032a6dfc2eba058a1fcc74bb8ef: Status 404 returned error can't find the container with id 1fcb384cb07b3486eafaabe6e7406da25f1e2032a6dfc2eba058a1fcc74bb8ef Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.371563 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-config" (OuterVolumeSpecName: "config") pod "11105f93-4bcd-4fd1-8640-ba20672c5df4" (UID: "11105f93-4bcd-4fd1-8640-ba20672c5df4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.403350 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-tn765"] Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.422396 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11105f93-4bcd-4fd1-8640-ba20672c5df4-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:37 crc kubenswrapper[4770]: W1004 03:20:37.442974 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4444038_6604_4e51_b27d_20d3616bf640.slice/crio-77c3c22bb2cdb8dbe4b9be8b7e236168e541c168357acc474291a61c021a0b13 WatchSource:0}: Error finding container 77c3c22bb2cdb8dbe4b9be8b7e236168e541c168357acc474291a61c021a0b13: Status 404 returned error can't find the container with id 77c3c22bb2cdb8dbe4b9be8b7e236168e541c168357acc474291a61c021a0b13 Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.457586 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-xfb5h"] Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.466260 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-xfb5h"] Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.690406 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11105f93-4bcd-4fd1-8640-ba20672c5df4" path="/var/lib/kubelet/pods/11105f93-4bcd-4fd1-8640-ba20672c5df4/volumes" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.730742 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 03:20:37 crc kubenswrapper[4770]: W1004 03:20:37.741313 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda36939a0_d10b_488e_8bd4_165b1f9ae4e9.slice/crio-32dcbadbecab7ac088279e37bd475ff16f530915676c670fa5c5005ecd755ba0 WatchSource:0}: Error finding container 32dcbadbecab7ac088279e37bd475ff16f530915676c670fa5c5005ecd755ba0: Status 404 returned error can't find the container with id 32dcbadbecab7ac088279e37bd475ff16f530915676c670fa5c5005ecd755ba0 Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.743475 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.743513 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.810312 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-lbkqb"] Oct 04 03:20:37 crc kubenswrapper[4770]: W1004 03:20:37.812422 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f5d1fda_e0a2_4fb4_8485_3b340a92c98f.slice/crio-47cfbe88f7ac51b3d428576333ed0187f9c3a0548c63f1fd5e6129c6a990e721 WatchSource:0}: Error finding container 47cfbe88f7ac51b3d428576333ed0187f9c3a0548c63f1fd5e6129c6a990e721: Status 404 returned error can't find the container with id 47cfbe88f7ac51b3d428576333ed0187f9c3a0548c63f1fd5e6129c6a990e721 Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.880461 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:37 crc kubenswrapper[4770]: I1004 03:20:37.880773 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:38 crc kubenswrapper[4770]: I1004 03:20:38.086660 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" event={"ID":"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f","Type":"ContainerStarted","Data":"47cfbe88f7ac51b3d428576333ed0187f9c3a0548c63f1fd5e6129c6a990e721"} Oct 04 03:20:38 crc kubenswrapper[4770]: I1004 03:20:38.087864 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a36939a0-d10b-488e-8bd4-165b1f9ae4e9","Type":"ContainerStarted","Data":"32dcbadbecab7ac088279e37bd475ff16f530915676c670fa5c5005ecd755ba0"} Oct 04 03:20:38 crc kubenswrapper[4770]: I1004 03:20:38.089625 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" event={"ID":"4b14dd83-328d-45dc-baa0-b63afe4e03a5","Type":"ContainerStarted","Data":"1fcb384cb07b3486eafaabe6e7406da25f1e2032a6dfc2eba058a1fcc74bb8ef"} Oct 04 03:20:38 crc kubenswrapper[4770]: I1004 03:20:38.092897 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-tn765" event={"ID":"b4444038-6604-4e51-b27d-20d3616bf640","Type":"ContainerStarted","Data":"77c3c22bb2cdb8dbe4b9be8b7e236168e541c168357acc474291a61c021a0b13"} Oct 04 03:20:38 crc kubenswrapper[4770]: I1004 03:20:38.222200 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 04 03:20:39 crc kubenswrapper[4770]: I1004 03:20:39.695397 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 03:20:39 crc kubenswrapper[4770]: I1004 03:20:39.809236 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-zjhc4"] Oct 04 03:20:39 crc kubenswrapper[4770]: I1004 03:20:39.891777 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-d5np4"] Oct 04 03:20:39 crc kubenswrapper[4770]: I1004 03:20:39.894562 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:39 crc kubenswrapper[4770]: I1004 03:20:39.919344 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-d5np4"] Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.011920 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-dns-svc\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.011969 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2rxh\" (UniqueName: \"kubernetes.io/projected/e31826ad-abb4-4996-ab4e-49185b5f88dd-kube-api-access-d2rxh\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.012050 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-nb\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.012146 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-sb\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.012169 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-config\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.114724 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-sb\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.114789 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-config\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.114843 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-dns-svc\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.114878 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2rxh\" (UniqueName: \"kubernetes.io/projected/e31826ad-abb4-4996-ab4e-49185b5f88dd-kube-api-access-d2rxh\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.114931 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-nb\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.115851 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-sb\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.115955 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-nb\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.116666 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-dns-svc\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.117111 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-config\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.134466 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2rxh\" (UniqueName: \"kubernetes.io/projected/e31826ad-abb4-4996-ab4e-49185b5f88dd-kube-api-access-d2rxh\") pod \"dnsmasq-dns-665d79f5c9-d5np4\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.239906 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.696928 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-d5np4"] Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.986825 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.992317 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.996968 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.996999 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.997317 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-rc4r2" Oct 04 03:20:40 crc kubenswrapper[4770]: I1004 03:20:40.998635 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.022137 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.129135 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.129558 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdfg5\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-kube-api-access-fdfg5\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.129716 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-cache\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.129794 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-lock\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.129939 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.133534 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" event={"ID":"e31826ad-abb4-4996-ab4e-49185b5f88dd","Type":"ContainerStarted","Data":"0abe1b17d87d19f76498cd5e9d09167c838803a9303bf34ba77feefd175b2041"} Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.231929 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: E1004 03:20:41.232110 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:20:41 crc kubenswrapper[4770]: E1004 03:20:41.232133 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:20:41 crc kubenswrapper[4770]: E1004 03:20:41.232184 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift podName:49d542d9-ee75-4f69-a103-0abb25d8c891 nodeName:}" failed. No retries permitted until 2025-10-04 03:20:41.732163844 +0000 UTC m=+1053.024173566 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift") pod "swift-storage-0" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891") : configmap "swift-ring-files" not found Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.232464 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdfg5\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-kube-api-access-fdfg5\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.232582 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-cache\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.232610 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-lock\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.232634 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.233023 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.233593 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-cache\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.233697 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-lock\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.254738 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdfg5\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-kube-api-access-fdfg5\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.260870 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.551152 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bl2rc"] Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.552760 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.556809 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.556818 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.556859 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.583709 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bl2rc"] Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.641260 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-dispersionconf\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.641625 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-scripts\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.641995 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-ring-data-devices\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.642205 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xgzk\" (UniqueName: \"kubernetes.io/projected/32226404-ae26-443d-aec7-766409996096-kube-api-access-9xgzk\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.642360 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32226404-ae26-443d-aec7-766409996096-etc-swift\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.642536 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-combined-ca-bundle\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.642741 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-swiftconf\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.745682 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-dispersionconf\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.746212 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-scripts\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.746308 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.746456 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-ring-data-devices\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: E1004 03:20:41.746738 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:20:41 crc kubenswrapper[4770]: E1004 03:20:41.746760 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:20:41 crc kubenswrapper[4770]: E1004 03:20:41.746823 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift podName:49d542d9-ee75-4f69-a103-0abb25d8c891 nodeName:}" failed. No retries permitted until 2025-10-04 03:20:42.746799604 +0000 UTC m=+1054.038809326 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift") pod "swift-storage-0" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891") : configmap "swift-ring-files" not found Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.747210 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-scripts\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.746535 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xgzk\" (UniqueName: \"kubernetes.io/projected/32226404-ae26-443d-aec7-766409996096-kube-api-access-9xgzk\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.748218 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32226404-ae26-443d-aec7-766409996096-etc-swift\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.748305 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-ring-data-devices\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.748415 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-combined-ca-bundle\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.748690 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-swiftconf\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.749330 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-dispersionconf\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.749370 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32226404-ae26-443d-aec7-766409996096-etc-swift\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.754627 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-combined-ca-bundle\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.756684 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-swiftconf\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.769977 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xgzk\" (UniqueName: \"kubernetes.io/projected/32226404-ae26-443d-aec7-766409996096-kube-api-access-9xgzk\") pod \"swift-ring-rebalance-bl2rc\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:41 crc kubenswrapper[4770]: I1004 03:20:41.915799 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:42 crc kubenswrapper[4770]: I1004 03:20:42.147432 4770 generic.go:334] "Generic (PLEG): container finished" podID="e31826ad-abb4-4996-ab4e-49185b5f88dd" containerID="eddfca7d848951d50e3fe2267ab70c8c527f62e474d01a2a10a096c3f9ffa905" exitCode=0 Oct 04 03:20:42 crc kubenswrapper[4770]: I1004 03:20:42.147598 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" event={"ID":"e31826ad-abb4-4996-ab4e-49185b5f88dd","Type":"ContainerDied","Data":"eddfca7d848951d50e3fe2267ab70c8c527f62e474d01a2a10a096c3f9ffa905"} Oct 04 03:20:42 crc kubenswrapper[4770]: I1004 03:20:42.189675 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bl2rc"] Oct 04 03:20:42 crc kubenswrapper[4770]: W1004 03:20:42.206982 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32226404_ae26_443d_aec7_766409996096.slice/crio-79f7e5d94d80bea77bfa535effa18b9090139257ad27662ebb02053ceacc563b WatchSource:0}: Error finding container 79f7e5d94d80bea77bfa535effa18b9090139257ad27662ebb02053ceacc563b: Status 404 returned error can't find the container with id 79f7e5d94d80bea77bfa535effa18b9090139257ad27662ebb02053ceacc563b Oct 04 03:20:42 crc kubenswrapper[4770]: I1004 03:20:42.774930 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:42 crc kubenswrapper[4770]: E1004 03:20:42.775634 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:20:42 crc kubenswrapper[4770]: E1004 03:20:42.775662 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:20:42 crc kubenswrapper[4770]: E1004 03:20:42.775736 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift podName:49d542d9-ee75-4f69-a103-0abb25d8c891 nodeName:}" failed. No retries permitted until 2025-10-04 03:20:44.775714475 +0000 UTC m=+1056.067724197 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift") pod "swift-storage-0" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891") : configmap "swift-ring-files" not found Oct 04 03:20:43 crc kubenswrapper[4770]: I1004 03:20:43.162882 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bl2rc" event={"ID":"32226404-ae26-443d-aec7-766409996096","Type":"ContainerStarted","Data":"79f7e5d94d80bea77bfa535effa18b9090139257ad27662ebb02053ceacc563b"} Oct 04 03:20:44 crc kubenswrapper[4770]: I1004 03:20:44.175093 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" event={"ID":"e31826ad-abb4-4996-ab4e-49185b5f88dd","Type":"ContainerStarted","Data":"caffb128695c98b74c1d1bbd2233e7ab75314b870bc7b1697cfec5bc0cbfcbbd"} Oct 04 03:20:44 crc kubenswrapper[4770]: I1004 03:20:44.175595 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:44 crc kubenswrapper[4770]: I1004 03:20:44.210513 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" podStartSLOduration=5.210487719 podStartE2EDuration="5.210487719s" podCreationTimestamp="2025-10-04 03:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:20:44.205261112 +0000 UTC m=+1055.497270834" watchObservedRunningTime="2025-10-04 03:20:44.210487719 +0000 UTC m=+1055.502497441" Oct 04 03:20:44 crc kubenswrapper[4770]: I1004 03:20:44.812475 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:44 crc kubenswrapper[4770]: E1004 03:20:44.814500 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:20:44 crc kubenswrapper[4770]: E1004 03:20:44.814535 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:20:44 crc kubenswrapper[4770]: E1004 03:20:44.814609 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift podName:49d542d9-ee75-4f69-a103-0abb25d8c891 nodeName:}" failed. No retries permitted until 2025-10-04 03:20:48.814588282 +0000 UTC m=+1060.106598084 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift") pod "swift-storage-0" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891") : configmap "swift-ring-files" not found Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.183951 4770 generic.go:334] "Generic (PLEG): container finished" podID="4b14dd83-328d-45dc-baa0-b63afe4e03a5" containerID="a5479f9eb0a1d6943d977d702010dc41db9d93288138e997b7db9e1efb1d6797" exitCode=0 Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.184126 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" event={"ID":"4b14dd83-328d-45dc-baa0-b63afe4e03a5","Type":"ContainerDied","Data":"a5479f9eb0a1d6943d977d702010dc41db9d93288138e997b7db9e1efb1d6797"} Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.186847 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-tn765" event={"ID":"b4444038-6604-4e51-b27d-20d3616bf640","Type":"ContainerStarted","Data":"3089621a8e69e00ac3e8841ec02e869f55aaa2a7703a475f407db0d736035cd6"} Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.189683 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" podUID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" containerName="dnsmasq-dns" containerID="cri-o://f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e" gracePeriod=10 Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.189968 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" event={"ID":"c81c7d43-13ea-4fb9-9f5e-f837e3959603","Type":"ContainerStarted","Data":"f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e"} Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.190020 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.194568 4770 generic.go:334] "Generic (PLEG): container finished" podID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" containerID="9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375" exitCode=0 Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.194829 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" event={"ID":"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f","Type":"ContainerDied","Data":"9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375"} Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.257573 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" podStartSLOduration=11.353745704 podStartE2EDuration="42.257557066s" podCreationTimestamp="2025-10-04 03:20:03 +0000 UTC" firstStartedPulling="2025-10-04 03:20:04.2635051 +0000 UTC m=+1015.555514812" lastFinishedPulling="2025-10-04 03:20:35.167316442 +0000 UTC m=+1046.459326174" observedRunningTime="2025-10-04 03:20:45.255579803 +0000 UTC m=+1056.547589515" watchObservedRunningTime="2025-10-04 03:20:45.257557066 +0000 UTC m=+1056.549566778" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.279036 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-tn765" podStartSLOduration=9.279017827 podStartE2EDuration="9.279017827s" podCreationTimestamp="2025-10-04 03:20:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:20:45.275211617 +0000 UTC m=+1056.567221349" watchObservedRunningTime="2025-10-04 03:20:45.279017827 +0000 UTC m=+1056.571027539" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.562258 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.626580 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-ovsdbserver-nb\") pod \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.626668 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-config\") pod \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.626770 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-dns-svc\") pod \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.626810 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dfg9\" (UniqueName: \"kubernetes.io/projected/4b14dd83-328d-45dc-baa0-b63afe4e03a5-kube-api-access-8dfg9\") pod \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\" (UID: \"4b14dd83-328d-45dc-baa0-b63afe4e03a5\") " Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.632502 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b14dd83-328d-45dc-baa0-b63afe4e03a5-kube-api-access-8dfg9" (OuterVolumeSpecName: "kube-api-access-8dfg9") pod "4b14dd83-328d-45dc-baa0-b63afe4e03a5" (UID: "4b14dd83-328d-45dc-baa0-b63afe4e03a5"). InnerVolumeSpecName "kube-api-access-8dfg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:45 crc kubenswrapper[4770]: E1004 03:20:45.649972 4770 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 04 03:20:45 crc kubenswrapper[4770]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 04 03:20:45 crc kubenswrapper[4770]: > podSandboxID="47cfbe88f7ac51b3d428576333ed0187f9c3a0548c63f1fd5e6129c6a990e721" Oct 04 03:20:45 crc kubenswrapper[4770]: E1004 03:20:45.651727 4770 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 04 03:20:45 crc kubenswrapper[4770]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n599h5cbh7ch5d4h66fh676hdbh546h95h88h5ffh55ch7fhch57ch687hddhc7h5fdh57dh674h56fh64ch98h9bh557h55dh646h54ch54fh5c4h597q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w4f29,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-8454ffc489-lbkqb_openstack(7f5d1fda-e0a2-4fb4-8485-3b340a92c98f): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 04 03:20:45 crc kubenswrapper[4770]: > logger="UnhandledError" Oct 04 03:20:45 crc kubenswrapper[4770]: E1004 03:20:45.654661 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" podUID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.657508 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4b14dd83-328d-45dc-baa0-b63afe4e03a5" (UID: "4b14dd83-328d-45dc-baa0-b63afe4e03a5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.664621 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-config" (OuterVolumeSpecName: "config") pod "4b14dd83-328d-45dc-baa0-b63afe4e03a5" (UID: "4b14dd83-328d-45dc-baa0-b63afe4e03a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.673421 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4b14dd83-328d-45dc-baa0-b63afe4e03a5" (UID: "4b14dd83-328d-45dc-baa0-b63afe4e03a5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.691835 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.728720 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dfg9\" (UniqueName: \"kubernetes.io/projected/4b14dd83-328d-45dc-baa0-b63afe4e03a5-kube-api-access-8dfg9\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.729161 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.729171 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.729180 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b14dd83-328d-45dc-baa0-b63afe4e03a5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.829750 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-config\") pod \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.829801 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-dns-svc\") pod \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.829831 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndc9m\" (UniqueName: \"kubernetes.io/projected/c81c7d43-13ea-4fb9-9f5e-f837e3959603-kube-api-access-ndc9m\") pod \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\" (UID: \"c81c7d43-13ea-4fb9-9f5e-f837e3959603\") " Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.839724 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c81c7d43-13ea-4fb9-9f5e-f837e3959603-kube-api-access-ndc9m" (OuterVolumeSpecName: "kube-api-access-ndc9m") pod "c81c7d43-13ea-4fb9-9f5e-f837e3959603" (UID: "c81c7d43-13ea-4fb9-9f5e-f837e3959603"). InnerVolumeSpecName "kube-api-access-ndc9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.888293 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-config" (OuterVolumeSpecName: "config") pod "c81c7d43-13ea-4fb9-9f5e-f837e3959603" (UID: "c81c7d43-13ea-4fb9-9f5e-f837e3959603"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.888474 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c81c7d43-13ea-4fb9-9f5e-f837e3959603" (UID: "c81c7d43-13ea-4fb9-9f5e-f837e3959603"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.931689 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.931738 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c81c7d43-13ea-4fb9-9f5e-f837e3959603-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:45 crc kubenswrapper[4770]: I1004 03:20:45.931753 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndc9m\" (UniqueName: \"kubernetes.io/projected/c81c7d43-13ea-4fb9-9f5e-f837e3959603-kube-api-access-ndc9m\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.105152 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.167981 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.207760 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a36939a0-d10b-488e-8bd4-165b1f9ae4e9","Type":"ContainerStarted","Data":"00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90"} Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.207807 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a36939a0-d10b-488e-8bd4-165b1f9ae4e9","Type":"ContainerStarted","Data":"38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed"} Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.207865 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.209744 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" event={"ID":"4b14dd83-328d-45dc-baa0-b63afe4e03a5","Type":"ContainerDied","Data":"1fcb384cb07b3486eafaabe6e7406da25f1e2032a6dfc2eba058a1fcc74bb8ef"} Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.209776 4770 scope.go:117] "RemoveContainer" containerID="a5479f9eb0a1d6943d977d702010dc41db9d93288138e997b7db9e1efb1d6797" Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.209902 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8749979c-zjhc4" Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.214055 4770 generic.go:334] "Generic (PLEG): container finished" podID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" containerID="f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e" exitCode=0 Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.215713 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.229203 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" event={"ID":"c81c7d43-13ea-4fb9-9f5e-f837e3959603","Type":"ContainerDied","Data":"f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e"} Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.229252 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-vgk4p" event={"ID":"c81c7d43-13ea-4fb9-9f5e-f837e3959603","Type":"ContainerDied","Data":"d271338295de3aa6f508507bec5f4bf458c936bb901478f304dbe351fe3fc9a7"} Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.231805 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.677641182 podStartE2EDuration="10.231787836s" podCreationTimestamp="2025-10-04 03:20:36 +0000 UTC" firstStartedPulling="2025-10-04 03:20:37.745435012 +0000 UTC m=+1049.037444724" lastFinishedPulling="2025-10-04 03:20:45.299581666 +0000 UTC m=+1056.591591378" observedRunningTime="2025-10-04 03:20:46.230477241 +0000 UTC m=+1057.522486943" watchObservedRunningTime="2025-10-04 03:20:46.231787836 +0000 UTC m=+1057.523797548" Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.288485 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-vgk4p"] Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.288909 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-vgk4p"] Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.313545 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-zjhc4"] Oct 04 03:20:46 crc kubenswrapper[4770]: I1004 03:20:46.355491 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b8749979c-zjhc4"] Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.689380 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b14dd83-328d-45dc-baa0-b63afe4e03a5" path="/var/lib/kubelet/pods/4b14dd83-328d-45dc-baa0-b63afe4e03a5/volumes" Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.690552 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" path="/var/lib/kubelet/pods/c81c7d43-13ea-4fb9-9f5e-f837e3959603/volumes" Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.881472 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-tqwql"] Oct 04 03:20:47 crc kubenswrapper[4770]: E1004 03:20:47.881844 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b14dd83-328d-45dc-baa0-b63afe4e03a5" containerName="init" Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.881866 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b14dd83-328d-45dc-baa0-b63afe4e03a5" containerName="init" Oct 04 03:20:47 crc kubenswrapper[4770]: E1004 03:20:47.881887 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" containerName="dnsmasq-dns" Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.881896 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" containerName="dnsmasq-dns" Oct 04 03:20:47 crc kubenswrapper[4770]: E1004 03:20:47.881914 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" containerName="init" Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.881923 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" containerName="init" Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.882145 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b14dd83-328d-45dc-baa0-b63afe4e03a5" containerName="init" Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.882162 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c81c7d43-13ea-4fb9-9f5e-f837e3959603" containerName="dnsmasq-dns" Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.882806 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tqwql" Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.894773 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-tqwql"] Oct 04 03:20:47 crc kubenswrapper[4770]: I1004 03:20:47.976571 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwshn\" (UniqueName: \"kubernetes.io/projected/263fdaa5-3f50-4f4e-8908-6974cee72f0d-kube-api-access-fwshn\") pod \"keystone-db-create-tqwql\" (UID: \"263fdaa5-3f50-4f4e-8908-6974cee72f0d\") " pod="openstack/keystone-db-create-tqwql" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.045239 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.078647 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwshn\" (UniqueName: \"kubernetes.io/projected/263fdaa5-3f50-4f4e-8908-6974cee72f0d-kube-api-access-fwshn\") pod \"keystone-db-create-tqwql\" (UID: \"263fdaa5-3f50-4f4e-8908-6974cee72f0d\") " pod="openstack/keystone-db-create-tqwql" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.102153 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwshn\" (UniqueName: \"kubernetes.io/projected/263fdaa5-3f50-4f4e-8908-6974cee72f0d-kube-api-access-fwshn\") pod \"keystone-db-create-tqwql\" (UID: \"263fdaa5-3f50-4f4e-8908-6974cee72f0d\") " pod="openstack/keystone-db-create-tqwql" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.107634 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.185710 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-dtz9z"] Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.186809 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dtz9z" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.201367 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tqwql" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.209181 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-dtz9z"] Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.286115 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdpzr\" (UniqueName: \"kubernetes.io/projected/6df7b787-d263-4732-a672-66a0b33c6369-kube-api-access-gdpzr\") pod \"placement-db-create-dtz9z\" (UID: \"6df7b787-d263-4732-a672-66a0b33c6369\") " pod="openstack/placement-db-create-dtz9z" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.391304 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdpzr\" (UniqueName: \"kubernetes.io/projected/6df7b787-d263-4732-a672-66a0b33c6369-kube-api-access-gdpzr\") pod \"placement-db-create-dtz9z\" (UID: \"6df7b787-d263-4732-a672-66a0b33c6369\") " pod="openstack/placement-db-create-dtz9z" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.397720 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-bgpjr"] Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.398991 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bgpjr" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.406813 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bgpjr"] Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.409263 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdpzr\" (UniqueName: \"kubernetes.io/projected/6df7b787-d263-4732-a672-66a0b33c6369-kube-api-access-gdpzr\") pod \"placement-db-create-dtz9z\" (UID: \"6df7b787-d263-4732-a672-66a0b33c6369\") " pod="openstack/placement-db-create-dtz9z" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.426663 4770 scope.go:117] "RemoveContainer" containerID="f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.493308 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfx5n\" (UniqueName: \"kubernetes.io/projected/4da6c58e-fb44-4acb-b49d-d29a0f82882f-kube-api-access-lfx5n\") pod \"glance-db-create-bgpjr\" (UID: \"4da6c58e-fb44-4acb-b49d-d29a0f82882f\") " pod="openstack/glance-db-create-bgpjr" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.494212 4770 scope.go:117] "RemoveContainer" containerID="e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.512025 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dtz9z" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.597155 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfx5n\" (UniqueName: \"kubernetes.io/projected/4da6c58e-fb44-4acb-b49d-d29a0f82882f-kube-api-access-lfx5n\") pod \"glance-db-create-bgpjr\" (UID: \"4da6c58e-fb44-4acb-b49d-d29a0f82882f\") " pod="openstack/glance-db-create-bgpjr" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.630810 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfx5n\" (UniqueName: \"kubernetes.io/projected/4da6c58e-fb44-4acb-b49d-d29a0f82882f-kube-api-access-lfx5n\") pod \"glance-db-create-bgpjr\" (UID: \"4da6c58e-fb44-4acb-b49d-d29a0f82882f\") " pod="openstack/glance-db-create-bgpjr" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.641572 4770 scope.go:117] "RemoveContainer" containerID="f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e" Oct 04 03:20:48 crc kubenswrapper[4770]: E1004 03:20:48.643350 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e\": container with ID starting with f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e not found: ID does not exist" containerID="f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.643393 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e"} err="failed to get container status \"f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e\": rpc error: code = NotFound desc = could not find container \"f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e\": container with ID starting with f16d146f81609d41ec62575a8f01079fdd20b4a418d655cac994c46d092a326e not found: ID does not exist" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.643421 4770 scope.go:117] "RemoveContainer" containerID="e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5" Oct 04 03:20:48 crc kubenswrapper[4770]: E1004 03:20:48.643842 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5\": container with ID starting with e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5 not found: ID does not exist" containerID="e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.643870 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5"} err="failed to get container status \"e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5\": rpc error: code = NotFound desc = could not find container \"e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5\": container with ID starting with e96ef1455a292fda6cd82c9682eea1bbe28536a5df040344a7299b967b10b2c5 not found: ID does not exist" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.769625 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-tqwql"] Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.898098 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bgpjr" Oct 04 03:20:48 crc kubenswrapper[4770]: I1004 03:20:48.905270 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:48 crc kubenswrapper[4770]: E1004 03:20:48.905524 4770 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:20:48 crc kubenswrapper[4770]: E1004 03:20:48.905560 4770 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:20:48 crc kubenswrapper[4770]: E1004 03:20:48.905639 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift podName:49d542d9-ee75-4f69-a103-0abb25d8c891 nodeName:}" failed. No retries permitted until 2025-10-04 03:20:56.90561435 +0000 UTC m=+1068.197624062 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift") pod "swift-storage-0" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891") : configmap "swift-ring-files" not found Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.037768 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-dtz9z"] Oct 04 03:20:49 crc kubenswrapper[4770]: W1004 03:20:49.037893 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6df7b787_d263_4732_a672_66a0b33c6369.slice/crio-d488a719730d3669afbf793a44c1cb6e0c2cd15989d09a4cb3b681aef0209f88 WatchSource:0}: Error finding container d488a719730d3669afbf793a44c1cb6e0c2cd15989d09a4cb3b681aef0209f88: Status 404 returned error can't find the container with id d488a719730d3669afbf793a44c1cb6e0c2cd15989d09a4cb3b681aef0209f88 Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.243811 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" event={"ID":"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f","Type":"ContainerStarted","Data":"2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69"} Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.244472 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.246442 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bl2rc" event={"ID":"32226404-ae26-443d-aec7-766409996096","Type":"ContainerStarted","Data":"2c5f1590663be7eb40c53863ac7c31baca0a8a6b4a313965c166156581e362a2"} Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.256043 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dtz9z" event={"ID":"6df7b787-d263-4732-a672-66a0b33c6369","Type":"ContainerStarted","Data":"d488a719730d3669afbf793a44c1cb6e0c2cd15989d09a4cb3b681aef0209f88"} Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.259998 4770 generic.go:334] "Generic (PLEG): container finished" podID="263fdaa5-3f50-4f4e-8908-6974cee72f0d" containerID="99abcb9a2d9a291ad93ba6f3aa28f7b9d7d5d9535d25d3f395d236f89a77c628" exitCode=0 Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.260085 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tqwql" event={"ID":"263fdaa5-3f50-4f4e-8908-6974cee72f0d","Type":"ContainerDied","Data":"99abcb9a2d9a291ad93ba6f3aa28f7b9d7d5d9535d25d3f395d236f89a77c628"} Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.260115 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tqwql" event={"ID":"263fdaa5-3f50-4f4e-8908-6974cee72f0d","Type":"ContainerStarted","Data":"e9df20a47c7a3424f1ef9aaa0f87794ec53de2b3f6cd23a758dbe333509d7ca8"} Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.269154 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" podStartSLOduration=13.269136175 podStartE2EDuration="13.269136175s" podCreationTimestamp="2025-10-04 03:20:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:20:49.266915017 +0000 UTC m=+1060.558924729" watchObservedRunningTime="2025-10-04 03:20:49.269136175 +0000 UTC m=+1060.561145887" Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.317766 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bl2rc" podStartSLOduration=2.028538742 podStartE2EDuration="8.317710027s" podCreationTimestamp="2025-10-04 03:20:41 +0000 UTC" firstStartedPulling="2025-10-04 03:20:42.211045715 +0000 UTC m=+1053.503055427" lastFinishedPulling="2025-10-04 03:20:48.500217 +0000 UTC m=+1059.792226712" observedRunningTime="2025-10-04 03:20:49.307386346 +0000 UTC m=+1060.599396068" watchObservedRunningTime="2025-10-04 03:20:49.317710027 +0000 UTC m=+1060.609719779" Oct 04 03:20:49 crc kubenswrapper[4770]: I1004 03:20:49.373081 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bgpjr"] Oct 04 03:20:49 crc kubenswrapper[4770]: W1004 03:20:49.425543 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4da6c58e_fb44_4acb_b49d_d29a0f82882f.slice/crio-335578b581e2db8050d6685230166b5956017e2739a6e3b30471bdc38dc5f963 WatchSource:0}: Error finding container 335578b581e2db8050d6685230166b5956017e2739a6e3b30471bdc38dc5f963: Status 404 returned error can't find the container with id 335578b581e2db8050d6685230166b5956017e2739a6e3b30471bdc38dc5f963 Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.241266 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.294996 4770 generic.go:334] "Generic (PLEG): container finished" podID="4da6c58e-fb44-4acb-b49d-d29a0f82882f" containerID="78a49b5bce516a915e4c40d15b73c189f051e0f39b9bb11033e9c1a9a3daf2b0" exitCode=0 Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.295071 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bgpjr" event={"ID":"4da6c58e-fb44-4acb-b49d-d29a0f82882f","Type":"ContainerDied","Data":"78a49b5bce516a915e4c40d15b73c189f051e0f39b9bb11033e9c1a9a3daf2b0"} Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.295096 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bgpjr" event={"ID":"4da6c58e-fb44-4acb-b49d-d29a0f82882f","Type":"ContainerStarted","Data":"335578b581e2db8050d6685230166b5956017e2739a6e3b30471bdc38dc5f963"} Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.298939 4770 generic.go:334] "Generic (PLEG): container finished" podID="6df7b787-d263-4732-a672-66a0b33c6369" containerID="2c9cd9b93490eaac4e08d158df7a1d401d8e49a8abea19fcf2d103c304bcc073" exitCode=0 Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.299091 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dtz9z" event={"ID":"6df7b787-d263-4732-a672-66a0b33c6369","Type":"ContainerDied","Data":"2c9cd9b93490eaac4e08d158df7a1d401d8e49a8abea19fcf2d103c304bcc073"} Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.340359 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-lbkqb"] Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.685804 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tqwql" Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.736178 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwshn\" (UniqueName: \"kubernetes.io/projected/263fdaa5-3f50-4f4e-8908-6974cee72f0d-kube-api-access-fwshn\") pod \"263fdaa5-3f50-4f4e-8908-6974cee72f0d\" (UID: \"263fdaa5-3f50-4f4e-8908-6974cee72f0d\") " Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.742172 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/263fdaa5-3f50-4f4e-8908-6974cee72f0d-kube-api-access-fwshn" (OuterVolumeSpecName: "kube-api-access-fwshn") pod "263fdaa5-3f50-4f4e-8908-6974cee72f0d" (UID: "263fdaa5-3f50-4f4e-8908-6974cee72f0d"). InnerVolumeSpecName "kube-api-access-fwshn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:50 crc kubenswrapper[4770]: I1004 03:20:50.838333 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwshn\" (UniqueName: \"kubernetes.io/projected/263fdaa5-3f50-4f4e-8908-6974cee72f0d-kube-api-access-fwshn\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.309417 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tqwql" event={"ID":"263fdaa5-3f50-4f4e-8908-6974cee72f0d","Type":"ContainerDied","Data":"e9df20a47c7a3424f1ef9aaa0f87794ec53de2b3f6cd23a758dbe333509d7ca8"} Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.309468 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9df20a47c7a3424f1ef9aaa0f87794ec53de2b3f6cd23a758dbe333509d7ca8" Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.309578 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tqwql" Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.310047 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" podUID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" containerName="dnsmasq-dns" containerID="cri-o://2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69" gracePeriod=10 Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.638935 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dtz9z" Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.767443 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdpzr\" (UniqueName: \"kubernetes.io/projected/6df7b787-d263-4732-a672-66a0b33c6369-kube-api-access-gdpzr\") pod \"6df7b787-d263-4732-a672-66a0b33c6369\" (UID: \"6df7b787-d263-4732-a672-66a0b33c6369\") " Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.775237 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df7b787-d263-4732-a672-66a0b33c6369-kube-api-access-gdpzr" (OuterVolumeSpecName: "kube-api-access-gdpzr") pod "6df7b787-d263-4732-a672-66a0b33c6369" (UID: "6df7b787-d263-4732-a672-66a0b33c6369"). InnerVolumeSpecName "kube-api-access-gdpzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.820933 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bgpjr" Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.828752 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.870420 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdpzr\" (UniqueName: \"kubernetes.io/projected/6df7b787-d263-4732-a672-66a0b33c6369-kube-api-access-gdpzr\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.971860 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-sb\") pod \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.971909 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-dns-svc\") pod \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.971962 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-nb\") pod \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.972059 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-config\") pod \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.972079 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4f29\" (UniqueName: \"kubernetes.io/projected/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-kube-api-access-w4f29\") pod \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\" (UID: \"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f\") " Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.972205 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfx5n\" (UniqueName: \"kubernetes.io/projected/4da6c58e-fb44-4acb-b49d-d29a0f82882f-kube-api-access-lfx5n\") pod \"4da6c58e-fb44-4acb-b49d-d29a0f82882f\" (UID: \"4da6c58e-fb44-4acb-b49d-d29a0f82882f\") " Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.976473 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4da6c58e-fb44-4acb-b49d-d29a0f82882f-kube-api-access-lfx5n" (OuterVolumeSpecName: "kube-api-access-lfx5n") pod "4da6c58e-fb44-4acb-b49d-d29a0f82882f" (UID: "4da6c58e-fb44-4acb-b49d-d29a0f82882f"). InnerVolumeSpecName "kube-api-access-lfx5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:51 crc kubenswrapper[4770]: I1004 03:20:51.979206 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-kube-api-access-w4f29" (OuterVolumeSpecName: "kube-api-access-w4f29") pod "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" (UID: "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f"). InnerVolumeSpecName "kube-api-access-w4f29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.024654 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" (UID: "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.025654 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" (UID: "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.034707 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" (UID: "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.053068 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-config" (OuterVolumeSpecName: "config") pod "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" (UID: "7f5d1fda-e0a2-4fb4-8485-3b340a92c98f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.074689 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.074956 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.075109 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.075232 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.075314 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4f29\" (UniqueName: \"kubernetes.io/projected/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f-kube-api-access-w4f29\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.075427 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfx5n\" (UniqueName: \"kubernetes.io/projected/4da6c58e-fb44-4acb-b49d-d29a0f82882f-kube-api-access-lfx5n\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.319545 4770 generic.go:334] "Generic (PLEG): container finished" podID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" containerID="2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69" exitCode=0 Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.319609 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" event={"ID":"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f","Type":"ContainerDied","Data":"2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69"} Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.319637 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" event={"ID":"7f5d1fda-e0a2-4fb4-8485-3b340a92c98f","Type":"ContainerDied","Data":"47cfbe88f7ac51b3d428576333ed0187f9c3a0548c63f1fd5e6129c6a990e721"} Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.319782 4770 scope.go:117] "RemoveContainer" containerID="2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.321148 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8454ffc489-lbkqb" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.323058 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bgpjr" event={"ID":"4da6c58e-fb44-4acb-b49d-d29a0f82882f","Type":"ContainerDied","Data":"335578b581e2db8050d6685230166b5956017e2739a6e3b30471bdc38dc5f963"} Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.323113 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="335578b581e2db8050d6685230166b5956017e2739a6e3b30471bdc38dc5f963" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.323241 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bgpjr" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.327190 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dtz9z" event={"ID":"6df7b787-d263-4732-a672-66a0b33c6369","Type":"ContainerDied","Data":"d488a719730d3669afbf793a44c1cb6e0c2cd15989d09a4cb3b681aef0209f88"} Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.327242 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d488a719730d3669afbf793a44c1cb6e0c2cd15989d09a4cb3b681aef0209f88" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.327296 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dtz9z" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.349272 4770 scope.go:117] "RemoveContainer" containerID="9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.384198 4770 scope.go:117] "RemoveContainer" containerID="2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69" Oct 04 03:20:52 crc kubenswrapper[4770]: E1004 03:20:52.385252 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69\": container with ID starting with 2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69 not found: ID does not exist" containerID="2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.385372 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69"} err="failed to get container status \"2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69\": rpc error: code = NotFound desc = could not find container \"2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69\": container with ID starting with 2556511a18c42b3754dc95882330aaba4cd7081d7d64a6a7e56a58b1d5454a69 not found: ID does not exist" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.385465 4770 scope.go:117] "RemoveContainer" containerID="9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375" Oct 04 03:20:52 crc kubenswrapper[4770]: E1004 03:20:52.386154 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375\": container with ID starting with 9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375 not found: ID does not exist" containerID="9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.386196 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375"} err="failed to get container status \"9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375\": rpc error: code = NotFound desc = could not find container \"9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375\": container with ID starting with 9293d8109c541e38977a73e43f9056c6a0e134a9125be569a88f45dcc8b83375 not found: ID does not exist" Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.391733 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-lbkqb"] Oct 04 03:20:52 crc kubenswrapper[4770]: I1004 03:20:52.403341 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8454ffc489-lbkqb"] Oct 04 03:20:53 crc kubenswrapper[4770]: I1004 03:20:53.684741 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" path="/var/lib/kubelet/pods/7f5d1fda-e0a2-4fb4-8485-3b340a92c98f/volumes" Oct 04 03:20:55 crc kubenswrapper[4770]: I1004 03:20:55.369893 4770 generic.go:334] "Generic (PLEG): container finished" podID="32226404-ae26-443d-aec7-766409996096" containerID="2c5f1590663be7eb40c53863ac7c31baca0a8a6b4a313965c166156581e362a2" exitCode=0 Oct 04 03:20:55 crc kubenswrapper[4770]: I1004 03:20:55.370000 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bl2rc" event={"ID":"32226404-ae26-443d-aec7-766409996096","Type":"ContainerDied","Data":"2c5f1590663be7eb40c53863ac7c31baca0a8a6b4a313965c166156581e362a2"} Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.388498 4770 generic.go:334] "Generic (PLEG): container finished" podID="2def82a8-5927-4ebe-ac87-e8ad232797ee" containerID="5305bb132085306589edf135617a3aa23523ea917b7840da309c820a5a534359" exitCode=0 Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.388639 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2def82a8-5927-4ebe-ac87-e8ad232797ee","Type":"ContainerDied","Data":"5305bb132085306589edf135617a3aa23523ea917b7840da309c820a5a534359"} Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.396280 4770 generic.go:334] "Generic (PLEG): container finished" podID="642fe06d-e5b8-461e-a1f7-dbcefc071945" containerID="0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01" exitCode=0 Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.396998 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"642fe06d-e5b8-461e-a1f7-dbcefc071945","Type":"ContainerDied","Data":"0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01"} Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.762357 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.792990 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-ring-data-devices\") pod \"32226404-ae26-443d-aec7-766409996096\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.793122 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-dispersionconf\") pod \"32226404-ae26-443d-aec7-766409996096\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.793171 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-scripts\") pod \"32226404-ae26-443d-aec7-766409996096\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.793254 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-combined-ca-bundle\") pod \"32226404-ae26-443d-aec7-766409996096\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.793282 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-swiftconf\") pod \"32226404-ae26-443d-aec7-766409996096\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.793330 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xgzk\" (UniqueName: \"kubernetes.io/projected/32226404-ae26-443d-aec7-766409996096-kube-api-access-9xgzk\") pod \"32226404-ae26-443d-aec7-766409996096\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.793412 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32226404-ae26-443d-aec7-766409996096-etc-swift\") pod \"32226404-ae26-443d-aec7-766409996096\" (UID: \"32226404-ae26-443d-aec7-766409996096\") " Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.794807 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "32226404-ae26-443d-aec7-766409996096" (UID: "32226404-ae26-443d-aec7-766409996096"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.794979 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32226404-ae26-443d-aec7-766409996096-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "32226404-ae26-443d-aec7-766409996096" (UID: "32226404-ae26-443d-aec7-766409996096"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.799260 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32226404-ae26-443d-aec7-766409996096-kube-api-access-9xgzk" (OuterVolumeSpecName: "kube-api-access-9xgzk") pod "32226404-ae26-443d-aec7-766409996096" (UID: "32226404-ae26-443d-aec7-766409996096"). InnerVolumeSpecName "kube-api-access-9xgzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.802582 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "32226404-ae26-443d-aec7-766409996096" (UID: "32226404-ae26-443d-aec7-766409996096"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.814241 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-scripts" (OuterVolumeSpecName: "scripts") pod "32226404-ae26-443d-aec7-766409996096" (UID: "32226404-ae26-443d-aec7-766409996096"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.817220 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "32226404-ae26-443d-aec7-766409996096" (UID: "32226404-ae26-443d-aec7-766409996096"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.822610 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32226404-ae26-443d-aec7-766409996096" (UID: "32226404-ae26-443d-aec7-766409996096"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.896731 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.896795 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.896812 4770 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.896826 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xgzk\" (UniqueName: \"kubernetes.io/projected/32226404-ae26-443d-aec7-766409996096-kube-api-access-9xgzk\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.896841 4770 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32226404-ae26-443d-aec7-766409996096-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.896854 4770 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32226404-ae26-443d-aec7-766409996096-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.896867 4770 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32226404-ae26-443d-aec7-766409996096-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:56 crc kubenswrapper[4770]: I1004 03:20:56.999476 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.006088 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") pod \"swift-storage-0\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " pod="openstack/swift-storage-0" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.212371 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.349720 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.410685 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"642fe06d-e5b8-461e-a1f7-dbcefc071945","Type":"ContainerStarted","Data":"4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749"} Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.411926 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.427345 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2def82a8-5927-4ebe-ac87-e8ad232797ee","Type":"ContainerStarted","Data":"19babfd2683be969879e3923b984d72aa3faecdef7166dfa931d52c111a1f74f"} Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.427918 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.436093 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bl2rc" event={"ID":"32226404-ae26-443d-aec7-766409996096","Type":"ContainerDied","Data":"79f7e5d94d80bea77bfa535effa18b9090139257ad27662ebb02053ceacc563b"} Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.436127 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79f7e5d94d80bea77bfa535effa18b9090139257ad27662ebb02053ceacc563b" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.436180 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bl2rc" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.498061 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.323810058 podStartE2EDuration="54.498024199s" podCreationTimestamp="2025-10-04 03:20:03 +0000 UTC" firstStartedPulling="2025-10-04 03:20:05.531255103 +0000 UTC m=+1016.823264805" lastFinishedPulling="2025-10-04 03:20:22.705469234 +0000 UTC m=+1033.997478946" observedRunningTime="2025-10-04 03:20:57.46180813 +0000 UTC m=+1068.753817862" watchObservedRunningTime="2025-10-04 03:20:57.498024199 +0000 UTC m=+1068.790033911" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.499031 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.595035813 podStartE2EDuration="55.499020865s" podCreationTimestamp="2025-10-04 03:20:02 +0000 UTC" firstStartedPulling="2025-10-04 03:20:04.833560571 +0000 UTC m=+1016.125570273" lastFinishedPulling="2025-10-04 03:20:22.737545623 +0000 UTC m=+1034.029555325" observedRunningTime="2025-10-04 03:20:57.49656524 +0000 UTC m=+1068.788574962" watchObservedRunningTime="2025-10-04 03:20:57.499020865 +0000 UTC m=+1068.791030587" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.918548 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f041-account-create-dt88f"] Oct 04 03:20:57 crc kubenswrapper[4770]: E1004 03:20:57.918955 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df7b787-d263-4732-a672-66a0b33c6369" containerName="mariadb-database-create" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.918972 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df7b787-d263-4732-a672-66a0b33c6369" containerName="mariadb-database-create" Oct 04 03:20:57 crc kubenswrapper[4770]: E1004 03:20:57.918986 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" containerName="init" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.918993 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" containerName="init" Oct 04 03:20:57 crc kubenswrapper[4770]: E1004 03:20:57.919118 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" containerName="dnsmasq-dns" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919132 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" containerName="dnsmasq-dns" Oct 04 03:20:57 crc kubenswrapper[4770]: E1004 03:20:57.919150 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4da6c58e-fb44-4acb-b49d-d29a0f82882f" containerName="mariadb-database-create" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919157 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4da6c58e-fb44-4acb-b49d-d29a0f82882f" containerName="mariadb-database-create" Oct 04 03:20:57 crc kubenswrapper[4770]: E1004 03:20:57.919173 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="263fdaa5-3f50-4f4e-8908-6974cee72f0d" containerName="mariadb-database-create" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919180 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="263fdaa5-3f50-4f4e-8908-6974cee72f0d" containerName="mariadb-database-create" Oct 04 03:20:57 crc kubenswrapper[4770]: E1004 03:20:57.919195 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32226404-ae26-443d-aec7-766409996096" containerName="swift-ring-rebalance" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919202 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="32226404-ae26-443d-aec7-766409996096" containerName="swift-ring-rebalance" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919359 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4da6c58e-fb44-4acb-b49d-d29a0f82882f" containerName="mariadb-database-create" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919373 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="263fdaa5-3f50-4f4e-8908-6974cee72f0d" containerName="mariadb-database-create" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919393 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df7b787-d263-4732-a672-66a0b33c6369" containerName="mariadb-database-create" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919404 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f5d1fda-e0a2-4fb4-8485-3b340a92c98f" containerName="dnsmasq-dns" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919412 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="32226404-ae26-443d-aec7-766409996096" containerName="swift-ring-rebalance" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.919924 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f041-account-create-dt88f" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.926296 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.932881 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f041-account-create-dt88f"] Oct 04 03:20:57 crc kubenswrapper[4770]: I1004 03:20:57.993237 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.020543 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c7ch\" (UniqueName: \"kubernetes.io/projected/19c6e72a-329c-4adc-8987-b227ff421e9e-kube-api-access-9c7ch\") pod \"keystone-f041-account-create-dt88f\" (UID: \"19c6e72a-329c-4adc-8987-b227ff421e9e\") " pod="openstack/keystone-f041-account-create-dt88f" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.122383 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c7ch\" (UniqueName: \"kubernetes.io/projected/19c6e72a-329c-4adc-8987-b227ff421e9e-kube-api-access-9c7ch\") pod \"keystone-f041-account-create-dt88f\" (UID: \"19c6e72a-329c-4adc-8987-b227ff421e9e\") " pod="openstack/keystone-f041-account-create-dt88f" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.157147 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c7ch\" (UniqueName: \"kubernetes.io/projected/19c6e72a-329c-4adc-8987-b227ff421e9e-kube-api-access-9c7ch\") pod \"keystone-f041-account-create-dt88f\" (UID: \"19c6e72a-329c-4adc-8987-b227ff421e9e\") " pod="openstack/keystone-f041-account-create-dt88f" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.205136 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ded7-account-create-cbnfn"] Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.206486 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ded7-account-create-cbnfn" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.212894 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.229510 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ded7-account-create-cbnfn"] Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.248073 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f041-account-create-dt88f" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.325195 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z897p\" (UniqueName: \"kubernetes.io/projected/d56b3add-faa1-4600-bdca-f4bf03a11250-kube-api-access-z897p\") pod \"placement-ded7-account-create-cbnfn\" (UID: \"d56b3add-faa1-4600-bdca-f4bf03a11250\") " pod="openstack/placement-ded7-account-create-cbnfn" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.426293 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z897p\" (UniqueName: \"kubernetes.io/projected/d56b3add-faa1-4600-bdca-f4bf03a11250-kube-api-access-z897p\") pod \"placement-ded7-account-create-cbnfn\" (UID: \"d56b3add-faa1-4600-bdca-f4bf03a11250\") " pod="openstack/placement-ded7-account-create-cbnfn" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.449708 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"8e860b016b1756b7cff523f827ad0837c9c9bd3312fbaa71d299a216caa7dc26"} Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.451972 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z897p\" (UniqueName: \"kubernetes.io/projected/d56b3add-faa1-4600-bdca-f4bf03a11250-kube-api-access-z897p\") pod \"placement-ded7-account-create-cbnfn\" (UID: \"d56b3add-faa1-4600-bdca-f4bf03a11250\") " pod="openstack/placement-ded7-account-create-cbnfn" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.495554 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-d8ca-account-create-kj8sq"] Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.496932 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d8ca-account-create-kj8sq" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.505751 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.511639 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d8ca-account-create-kj8sq"] Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.536572 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ded7-account-create-cbnfn" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.629243 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpk26\" (UniqueName: \"kubernetes.io/projected/6159c490-e51a-4a0b-a570-35f9bd8b82c8-kube-api-access-wpk26\") pod \"glance-d8ca-account-create-kj8sq\" (UID: \"6159c490-e51a-4a0b-a570-35f9bd8b82c8\") " pod="openstack/glance-d8ca-account-create-kj8sq" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.713697 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f041-account-create-dt88f"] Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.731963 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpk26\" (UniqueName: \"kubernetes.io/projected/6159c490-e51a-4a0b-a570-35f9bd8b82c8-kube-api-access-wpk26\") pod \"glance-d8ca-account-create-kj8sq\" (UID: \"6159c490-e51a-4a0b-a570-35f9bd8b82c8\") " pod="openstack/glance-d8ca-account-create-kj8sq" Oct 04 03:20:58 crc kubenswrapper[4770]: W1004 03:20:58.739136 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19c6e72a_329c_4adc_8987_b227ff421e9e.slice/crio-8e689b592b688720563ef56ec8fe181dfbe40ce2f6ded9a3004ccf46afc1564f WatchSource:0}: Error finding container 8e689b592b688720563ef56ec8fe181dfbe40ce2f6ded9a3004ccf46afc1564f: Status 404 returned error can't find the container with id 8e689b592b688720563ef56ec8fe181dfbe40ce2f6ded9a3004ccf46afc1564f Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.754326 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpk26\" (UniqueName: \"kubernetes.io/projected/6159c490-e51a-4a0b-a570-35f9bd8b82c8-kube-api-access-wpk26\") pod \"glance-d8ca-account-create-kj8sq\" (UID: \"6159c490-e51a-4a0b-a570-35f9bd8b82c8\") " pod="openstack/glance-d8ca-account-create-kj8sq" Oct 04 03:20:58 crc kubenswrapper[4770]: I1004 03:20:58.826458 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d8ca-account-create-kj8sq" Oct 04 03:20:59 crc kubenswrapper[4770]: I1004 03:20:59.014633 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ded7-account-create-cbnfn"] Oct 04 03:20:59 crc kubenswrapper[4770]: I1004 03:20:59.478816 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ded7-account-create-cbnfn" event={"ID":"d56b3add-faa1-4600-bdca-f4bf03a11250","Type":"ContainerStarted","Data":"1cb8b0950182c8559e48dc79318dc993602cadf6599a8a91e064f0266f789dda"} Oct 04 03:20:59 crc kubenswrapper[4770]: I1004 03:20:59.479248 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ded7-account-create-cbnfn" event={"ID":"d56b3add-faa1-4600-bdca-f4bf03a11250","Type":"ContainerStarted","Data":"75fc1ef3c0352c1bf4001d9f7939dda7183993c0b550198decc02f508366ef7c"} Oct 04 03:20:59 crc kubenswrapper[4770]: I1004 03:20:59.481538 4770 generic.go:334] "Generic (PLEG): container finished" podID="19c6e72a-329c-4adc-8987-b227ff421e9e" containerID="3ce2dd1c186025121805ddd2cc61bd1c00a3f693c6b3053acc59af366a0c89f3" exitCode=0 Oct 04 03:20:59 crc kubenswrapper[4770]: I1004 03:20:59.481608 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f041-account-create-dt88f" event={"ID":"19c6e72a-329c-4adc-8987-b227ff421e9e","Type":"ContainerDied","Data":"3ce2dd1c186025121805ddd2cc61bd1c00a3f693c6b3053acc59af366a0c89f3"} Oct 04 03:20:59 crc kubenswrapper[4770]: I1004 03:20:59.481631 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f041-account-create-dt88f" event={"ID":"19c6e72a-329c-4adc-8987-b227ff421e9e","Type":"ContainerStarted","Data":"8e689b592b688720563ef56ec8fe181dfbe40ce2f6ded9a3004ccf46afc1564f"} Oct 04 03:20:59 crc kubenswrapper[4770]: I1004 03:20:59.494506 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"f824e75e4b4bc8b5e89c5df2fd7e73c49c2760133924bc91336e17d29f37ffa6"} Oct 04 03:20:59 crc kubenswrapper[4770]: I1004 03:20:59.502643 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-ded7-account-create-cbnfn" podStartSLOduration=1.502618587 podStartE2EDuration="1.502618587s" podCreationTimestamp="2025-10-04 03:20:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:20:59.494863133 +0000 UTC m=+1070.786872865" watchObservedRunningTime="2025-10-04 03:20:59.502618587 +0000 UTC m=+1070.794628319" Oct 04 03:20:59 crc kubenswrapper[4770]: I1004 03:20:59.593065 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d8ca-account-create-kj8sq"] Oct 04 03:20:59 crc kubenswrapper[4770]: W1004 03:20:59.599110 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6159c490_e51a_4a0b_a570_35f9bd8b82c8.slice/crio-f248f02ca53eff27b47c3c7dab519134ae1292bdf7247790a714589eb143587a WatchSource:0}: Error finding container f248f02ca53eff27b47c3c7dab519134ae1292bdf7247790a714589eb143587a: Status 404 returned error can't find the container with id f248f02ca53eff27b47c3c7dab519134ae1292bdf7247790a714589eb143587a Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.508066 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"8f1dd12bc897fdadff2fb989c2bead87e41df04b90e728e0cf683bc33de025c4"} Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.508595 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"290870635c010a1c742ef38d88d7a95bc4bda2febe2172b79b9d8b8044db9b56"} Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.508616 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"804fe60a8b7c232d84bf3cd9ed92ad47513e9d616208774db899e1a670a8fa9d"} Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.511298 4770 generic.go:334] "Generic (PLEG): container finished" podID="d56b3add-faa1-4600-bdca-f4bf03a11250" containerID="1cb8b0950182c8559e48dc79318dc993602cadf6599a8a91e064f0266f789dda" exitCode=0 Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.511382 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ded7-account-create-cbnfn" event={"ID":"d56b3add-faa1-4600-bdca-f4bf03a11250","Type":"ContainerDied","Data":"1cb8b0950182c8559e48dc79318dc993602cadf6599a8a91e064f0266f789dda"} Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.513782 4770 generic.go:334] "Generic (PLEG): container finished" podID="6159c490-e51a-4a0b-a570-35f9bd8b82c8" containerID="6ca4790905dd98d2b3d8d0601e85120a413fe7048ebdf88f7712f01b13767f88" exitCode=0 Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.514048 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d8ca-account-create-kj8sq" event={"ID":"6159c490-e51a-4a0b-a570-35f9bd8b82c8","Type":"ContainerDied","Data":"6ca4790905dd98d2b3d8d0601e85120a413fe7048ebdf88f7712f01b13767f88"} Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.514078 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d8ca-account-create-kj8sq" event={"ID":"6159c490-e51a-4a0b-a570-35f9bd8b82c8","Type":"ContainerStarted","Data":"f248f02ca53eff27b47c3c7dab519134ae1292bdf7247790a714589eb143587a"} Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.850202 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f041-account-create-dt88f" Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.980487 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c7ch\" (UniqueName: \"kubernetes.io/projected/19c6e72a-329c-4adc-8987-b227ff421e9e-kube-api-access-9c7ch\") pod \"19c6e72a-329c-4adc-8987-b227ff421e9e\" (UID: \"19c6e72a-329c-4adc-8987-b227ff421e9e\") " Oct 04 03:21:00 crc kubenswrapper[4770]: I1004 03:21:00.988320 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19c6e72a-329c-4adc-8987-b227ff421e9e-kube-api-access-9c7ch" (OuterVolumeSpecName: "kube-api-access-9c7ch") pod "19c6e72a-329c-4adc-8987-b227ff421e9e" (UID: "19c6e72a-329c-4adc-8987-b227ff421e9e"). InnerVolumeSpecName "kube-api-access-9c7ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.082768 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c7ch\" (UniqueName: \"kubernetes.io/projected/19c6e72a-329c-4adc-8987-b227ff421e9e-kube-api-access-9c7ch\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.532078 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"013814fda0f11cb6869f76d4caabd04fe6d67c24b613c2843c512e76f378e1d2"} Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.535225 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f041-account-create-dt88f" Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.535222 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f041-account-create-dt88f" event={"ID":"19c6e72a-329c-4adc-8987-b227ff421e9e","Type":"ContainerDied","Data":"8e689b592b688720563ef56ec8fe181dfbe40ce2f6ded9a3004ccf46afc1564f"} Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.535321 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e689b592b688720563ef56ec8fe181dfbe40ce2f6ded9a3004ccf46afc1564f" Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.795614 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.795718 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.795811 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.797118 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"73d28cffdd4e71105ae4d9aa35d59b331f61e307981b18b94799778a5b0a5a8d"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:21:01 crc kubenswrapper[4770]: I1004 03:21:01.797229 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://73d28cffdd4e71105ae4d9aa35d59b331f61e307981b18b94799778a5b0a5a8d" gracePeriod=600 Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.002765 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ded7-account-create-cbnfn" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.114647 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z897p\" (UniqueName: \"kubernetes.io/projected/d56b3add-faa1-4600-bdca-f4bf03a11250-kube-api-access-z897p\") pod \"d56b3add-faa1-4600-bdca-f4bf03a11250\" (UID: \"d56b3add-faa1-4600-bdca-f4bf03a11250\") " Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.121573 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d8ca-account-create-kj8sq" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.126321 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d56b3add-faa1-4600-bdca-f4bf03a11250-kube-api-access-z897p" (OuterVolumeSpecName: "kube-api-access-z897p") pod "d56b3add-faa1-4600-bdca-f4bf03a11250" (UID: "d56b3add-faa1-4600-bdca-f4bf03a11250"). InnerVolumeSpecName "kube-api-access-z897p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.220033 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpk26\" (UniqueName: \"kubernetes.io/projected/6159c490-e51a-4a0b-a570-35f9bd8b82c8-kube-api-access-wpk26\") pod \"6159c490-e51a-4a0b-a570-35f9bd8b82c8\" (UID: \"6159c490-e51a-4a0b-a570-35f9bd8b82c8\") " Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.220699 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z897p\" (UniqueName: \"kubernetes.io/projected/d56b3add-faa1-4600-bdca-f4bf03a11250-kube-api-access-z897p\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.224469 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6159c490-e51a-4a0b-a570-35f9bd8b82c8-kube-api-access-wpk26" (OuterVolumeSpecName: "kube-api-access-wpk26") pod "6159c490-e51a-4a0b-a570-35f9bd8b82c8" (UID: "6159c490-e51a-4a0b-a570-35f9bd8b82c8"). InnerVolumeSpecName "kube-api-access-wpk26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.321895 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpk26\" (UniqueName: \"kubernetes.io/projected/6159c490-e51a-4a0b-a570-35f9bd8b82c8-kube-api-access-wpk26\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.557647 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"0bbc28ed4f22995dd014ebc4247f63d2b41fd2841242ce9db1f8c81ed20dd754"} Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.557691 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"62177bc6b3f5210e61d4fa109ad0edab21794e6c92a8e65133e7ae6f064ed8f9"} Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.557702 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"7470ca0d05ae46b498189734f3f7a811e3f808a06e71678312140fd5f7c389ea"} Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.560293 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ded7-account-create-cbnfn" event={"ID":"d56b3add-faa1-4600-bdca-f4bf03a11250","Type":"ContainerDied","Data":"75fc1ef3c0352c1bf4001d9f7939dda7183993c0b550198decc02f508366ef7c"} Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.560330 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75fc1ef3c0352c1bf4001d9f7939dda7183993c0b550198decc02f508366ef7c" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.560381 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ded7-account-create-cbnfn" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.581210 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="73d28cffdd4e71105ae4d9aa35d59b331f61e307981b18b94799778a5b0a5a8d" exitCode=0 Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.581322 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"73d28cffdd4e71105ae4d9aa35d59b331f61e307981b18b94799778a5b0a5a8d"} Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.581357 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"d6b0b7a6fef39110ff7706ef744d6f617ed372a11387e2977774d987a4194d8f"} Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.581378 4770 scope.go:117] "RemoveContainer" containerID="1f8ff4d68e47173d152e933c25298fdeacbf8df65acd72aab55c2b5b7e5c9c52" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.609558 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d8ca-account-create-kj8sq" event={"ID":"6159c490-e51a-4a0b-a570-35f9bd8b82c8","Type":"ContainerDied","Data":"f248f02ca53eff27b47c3c7dab519134ae1292bdf7247790a714589eb143587a"} Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.609603 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f248f02ca53eff27b47c3c7dab519134ae1292bdf7247790a714589eb143587a" Oct 04 03:21:02 crc kubenswrapper[4770]: I1004 03:21:02.609659 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d8ca-account-create-kj8sq" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.586746 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-ql7kt"] Oct 04 03:21:03 crc kubenswrapper[4770]: E1004 03:21:03.587607 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56b3add-faa1-4600-bdca-f4bf03a11250" containerName="mariadb-account-create" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.587621 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56b3add-faa1-4600-bdca-f4bf03a11250" containerName="mariadb-account-create" Oct 04 03:21:03 crc kubenswrapper[4770]: E1004 03:21:03.587640 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6159c490-e51a-4a0b-a570-35f9bd8b82c8" containerName="mariadb-account-create" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.587645 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6159c490-e51a-4a0b-a570-35f9bd8b82c8" containerName="mariadb-account-create" Oct 04 03:21:03 crc kubenswrapper[4770]: E1004 03:21:03.587660 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19c6e72a-329c-4adc-8987-b227ff421e9e" containerName="mariadb-account-create" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.587667 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="19c6e72a-329c-4adc-8987-b227ff421e9e" containerName="mariadb-account-create" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.587822 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6159c490-e51a-4a0b-a570-35f9bd8b82c8" containerName="mariadb-account-create" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.587858 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="19c6e72a-329c-4adc-8987-b227ff421e9e" containerName="mariadb-account-create" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.587873 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d56b3add-faa1-4600-bdca-f4bf03a11250" containerName="mariadb-account-create" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.588437 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.598497 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ql7kt"] Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.599730 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.600276 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mc7pb" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.612561 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-c2v9x" podUID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" containerName="ovn-controller" probeResult="failure" output=< Oct 04 03:21:03 crc kubenswrapper[4770]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 03:21:03 crc kubenswrapper[4770]: > Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.645825 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.651736 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.747995 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-combined-ca-bundle\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.748362 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-config-data\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.748406 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2cpr\" (UniqueName: \"kubernetes.io/projected/76095968-a5a2-4a39-bf86-5faafb9d76d3-kube-api-access-h2cpr\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.748456 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-db-sync-config-data\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.850979 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-combined-ca-bundle\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.851125 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-config-data\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.851155 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2cpr\" (UniqueName: \"kubernetes.io/projected/76095968-a5a2-4a39-bf86-5faafb9d76d3-kube-api-access-h2cpr\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.851198 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-db-sync-config-data\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.858982 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-db-sync-config-data\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.860866 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-config-data\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.863022 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-combined-ca-bundle\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.872821 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-c2v9x-config-rv799"] Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.873988 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.876128 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.880569 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2cpr\" (UniqueName: \"kubernetes.io/projected/76095968-a5a2-4a39-bf86-5faafb9d76d3-kube-api-access-h2cpr\") pod \"glance-db-sync-ql7kt\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.881977 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-c2v9x-config-rv799"] Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.919133 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.952894 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run-ovn\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.952944 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtggq\" (UniqueName: \"kubernetes.io/projected/31252013-1c9c-4586-a45b-e693a1149181-kube-api-access-xtggq\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.952984 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.953053 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-scripts\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.953118 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-log-ovn\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:03 crc kubenswrapper[4770]: I1004 03:21:03.953134 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-additional-scripts\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.055242 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-log-ovn\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.055285 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-additional-scripts\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.055318 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run-ovn\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.055346 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtggq\" (UniqueName: \"kubernetes.io/projected/31252013-1c9c-4586-a45b-e693a1149181-kube-api-access-xtggq\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.055385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.055414 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-scripts\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.056510 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-log-ovn\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.057049 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run-ovn\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.057507 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-additional-scripts\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.060636 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-scripts\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.060689 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.086066 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtggq\" (UniqueName: \"kubernetes.io/projected/31252013-1c9c-4586-a45b-e693a1149181-kube-api-access-xtggq\") pod \"ovn-controller-c2v9x-config-rv799\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.262718 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.641946 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"4da7895e64c6f0a051ead34f62c7e31342e26f449bd134c7e3529d04b73c2f3d"} Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.642782 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"6c9a46ab54e3016848961bed1d5d1cc0eeddcff8edf1e007687b96267338090d"} Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.642800 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"336ce838d7770e81da4c7557a31fa795b90be60a48b3e9b579e1d82a0a17207a"} Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.704259 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ql7kt"] Oct 04 03:21:04 crc kubenswrapper[4770]: W1004 03:21:04.721144 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76095968_a5a2_4a39_bf86_5faafb9d76d3.slice/crio-e3abcaa6dd73d67eb01e1e9b20df49fdbaa75c4b93f855918976362a19c494d4 WatchSource:0}: Error finding container e3abcaa6dd73d67eb01e1e9b20df49fdbaa75c4b93f855918976362a19c494d4: Status 404 returned error can't find the container with id e3abcaa6dd73d67eb01e1e9b20df49fdbaa75c4b93f855918976362a19c494d4 Oct 04 03:21:04 crc kubenswrapper[4770]: I1004 03:21:04.768593 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-c2v9x-config-rv799"] Oct 04 03:21:04 crc kubenswrapper[4770]: W1004 03:21:04.778597 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31252013_1c9c_4586_a45b_e693a1149181.slice/crio-a4b5696ad6470d0b2c262d00e6ab5b12ba86a319e893ae2d4d1424ec8de69c8d WatchSource:0}: Error finding container a4b5696ad6470d0b2c262d00e6ab5b12ba86a319e893ae2d4d1424ec8de69c8d: Status 404 returned error can't find the container with id a4b5696ad6470d0b2c262d00e6ab5b12ba86a319e893ae2d4d1424ec8de69c8d Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.653876 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"345b8c0092ebcc04825e922384f0c8ee2e508d97ba89f66363c61e92ce3d2670"} Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.654629 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"2aea2997649b938d2836a8f46cd25d7caab6d10e8f2aeeac5509198aa88b89ca"} Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.654646 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"6385e85337f2e8e8a6703c8fe588bbfb6105b68f63512703a1ed8a03cd155903"} Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.654656 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerStarted","Data":"72b2d91a64a79c90616f09e24250e6b65a49ae69602fb773330ae5f11fcb04d4"} Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.657788 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ql7kt" event={"ID":"76095968-a5a2-4a39-bf86-5faafb9d76d3","Type":"ContainerStarted","Data":"e3abcaa6dd73d67eb01e1e9b20df49fdbaa75c4b93f855918976362a19c494d4"} Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.659632 4770 generic.go:334] "Generic (PLEG): container finished" podID="31252013-1c9c-4586-a45b-e693a1149181" containerID="f386f612a27ebba4fbd5c97ad2193f2f49abf1418ef36ff9c6115064a41b5ab6" exitCode=0 Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.659665 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x-config-rv799" event={"ID":"31252013-1c9c-4586-a45b-e693a1149181","Type":"ContainerDied","Data":"f386f612a27ebba4fbd5c97ad2193f2f49abf1418ef36ff9c6115064a41b5ab6"} Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.659699 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x-config-rv799" event={"ID":"31252013-1c9c-4586-a45b-e693a1149181","Type":"ContainerStarted","Data":"a4b5696ad6470d0b2c262d00e6ab5b12ba86a319e893ae2d4d1424ec8de69c8d"} Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.729245 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.925334981 podStartE2EDuration="26.729201082s" podCreationTimestamp="2025-10-04 03:20:39 +0000 UTC" firstStartedPulling="2025-10-04 03:20:57.966989073 +0000 UTC m=+1069.258998795" lastFinishedPulling="2025-10-04 03:21:03.770855184 +0000 UTC m=+1075.062864896" observedRunningTime="2025-10-04 03:21:05.702176025 +0000 UTC m=+1076.994185737" watchObservedRunningTime="2025-10-04 03:21:05.729201082 +0000 UTC m=+1077.021210814" Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.950594 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77594f5649-9f4f4"] Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.955695 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.960621 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 04 03:21:05 crc kubenswrapper[4770]: I1004 03:21:05.969179 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-9f4f4"] Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.124695 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-svc\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.124797 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4jfg\" (UniqueName: \"kubernetes.io/projected/03359d36-a52b-4969-b1c6-38eaf57d89fa-kube-api-access-k4jfg\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.124893 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-config\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.124964 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-swift-storage-0\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.125123 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-sb\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.125143 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-nb\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.226412 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-sb\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.226811 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-nb\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.226861 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-svc\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.226927 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4jfg\" (UniqueName: \"kubernetes.io/projected/03359d36-a52b-4969-b1c6-38eaf57d89fa-kube-api-access-k4jfg\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.226959 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-config\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.226981 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-swift-storage-0\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.228150 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-swift-storage-0\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.229671 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-sb\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.229687 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-nb\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.237590 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-config\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.238092 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-svc\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.251817 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4jfg\" (UniqueName: \"kubernetes.io/projected/03359d36-a52b-4969-b1c6-38eaf57d89fa-kube-api-access-k4jfg\") pod \"dnsmasq-dns-77594f5649-9f4f4\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.282459 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.759792 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-9f4f4"] Oct 04 03:21:06 crc kubenswrapper[4770]: I1004 03:21:06.910105 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.047312 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-additional-scripts\") pod \"31252013-1c9c-4586-a45b-e693a1149181\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.047347 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-log-ovn\") pod \"31252013-1c9c-4586-a45b-e693a1149181\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.047381 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-scripts\") pod \"31252013-1c9c-4586-a45b-e693a1149181\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.047542 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtggq\" (UniqueName: \"kubernetes.io/projected/31252013-1c9c-4586-a45b-e693a1149181-kube-api-access-xtggq\") pod \"31252013-1c9c-4586-a45b-e693a1149181\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.047575 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run-ovn\") pod \"31252013-1c9c-4586-a45b-e693a1149181\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.047596 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run\") pod \"31252013-1c9c-4586-a45b-e693a1149181\" (UID: \"31252013-1c9c-4586-a45b-e693a1149181\") " Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.047885 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run" (OuterVolumeSpecName: "var-run") pod "31252013-1c9c-4586-a45b-e693a1149181" (UID: "31252013-1c9c-4586-a45b-e693a1149181"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.048296 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "31252013-1c9c-4586-a45b-e693a1149181" (UID: "31252013-1c9c-4586-a45b-e693a1149181"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.048459 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "31252013-1c9c-4586-a45b-e693a1149181" (UID: "31252013-1c9c-4586-a45b-e693a1149181"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.049476 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "31252013-1c9c-4586-a45b-e693a1149181" (UID: "31252013-1c9c-4586-a45b-e693a1149181"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.049608 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-scripts" (OuterVolumeSpecName: "scripts") pod "31252013-1c9c-4586-a45b-e693a1149181" (UID: "31252013-1c9c-4586-a45b-e693a1149181"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.054626 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31252013-1c9c-4586-a45b-e693a1149181-kube-api-access-xtggq" (OuterVolumeSpecName: "kube-api-access-xtggq") pod "31252013-1c9c-4586-a45b-e693a1149181" (UID: "31252013-1c9c-4586-a45b-e693a1149181"). InnerVolumeSpecName "kube-api-access-xtggq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.149303 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtggq\" (UniqueName: \"kubernetes.io/projected/31252013-1c9c-4586-a45b-e693a1149181-kube-api-access-xtggq\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.149333 4770 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.149344 4770 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.149352 4770 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.149362 4770 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/31252013-1c9c-4586-a45b-e693a1149181-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.149369 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31252013-1c9c-4586-a45b-e693a1149181-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.678080 4770 generic.go:334] "Generic (PLEG): container finished" podID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerID="cdfe922c4750454f542d0e693ca4bec380e1bd0655e34c774bacd5179bdeae43" exitCode=0 Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.690473 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x-config-rv799" Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.698822 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" event={"ID":"03359d36-a52b-4969-b1c6-38eaf57d89fa","Type":"ContainerDied","Data":"cdfe922c4750454f542d0e693ca4bec380e1bd0655e34c774bacd5179bdeae43"} Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.698898 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" event={"ID":"03359d36-a52b-4969-b1c6-38eaf57d89fa","Type":"ContainerStarted","Data":"fb7b8cc9baf0021ac1683dd10800180be95ceb9590fbde17ce68b45b48cd6e83"} Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.698925 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x-config-rv799" event={"ID":"31252013-1c9c-4586-a45b-e693a1149181","Type":"ContainerDied","Data":"a4b5696ad6470d0b2c262d00e6ab5b12ba86a319e893ae2d4d1424ec8de69c8d"} Oct 04 03:21:07 crc kubenswrapper[4770]: I1004 03:21:07.698953 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4b5696ad6470d0b2c262d00e6ab5b12ba86a319e893ae2d4d1424ec8de69c8d" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.001301 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-c2v9x-config-rv799"] Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.007487 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-c2v9x-config-rv799"] Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.127579 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-c2v9x-config-2pzlq"] Oct 04 03:21:08 crc kubenswrapper[4770]: E1004 03:21:08.127944 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31252013-1c9c-4586-a45b-e693a1149181" containerName="ovn-config" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.127964 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="31252013-1c9c-4586-a45b-e693a1149181" containerName="ovn-config" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.129618 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="31252013-1c9c-4586-a45b-e693a1149181" containerName="ovn-config" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.130322 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.132101 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.139942 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-c2v9x-config-2pzlq"] Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.278444 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-scripts\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.278833 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-log-ovn\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.278901 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-additional-scripts\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.278941 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf4r8\" (UniqueName: \"kubernetes.io/projected/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-kube-api-access-pf4r8\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.278999 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.279053 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run-ovn\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.380708 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-scripts\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.380768 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-log-ovn\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.380811 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-additional-scripts\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.380832 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf4r8\" (UniqueName: \"kubernetes.io/projected/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-kube-api-access-pf4r8\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.380879 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.380935 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run-ovn\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.381484 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-log-ovn\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.381563 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run-ovn\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.381584 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.382254 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-additional-scripts\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.383687 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-scripts\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.410933 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf4r8\" (UniqueName: \"kubernetes.io/projected/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-kube-api-access-pf4r8\") pod \"ovn-controller-c2v9x-config-2pzlq\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.450900 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.616598 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-c2v9x" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.705989 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" event={"ID":"03359d36-a52b-4969-b1c6-38eaf57d89fa","Type":"ContainerStarted","Data":"95c4469d744bdeafad6ffb6c0b084279944cb12985e2d21884cf1967cf113de9"} Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.706275 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.738058 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" podStartSLOduration=3.738039086 podStartE2EDuration="3.738039086s" podCreationTimestamp="2025-10-04 03:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:21:08.727533051 +0000 UTC m=+1080.019542763" watchObservedRunningTime="2025-10-04 03:21:08.738039086 +0000 UTC m=+1080.030048798" Oct 04 03:21:08 crc kubenswrapper[4770]: I1004 03:21:08.891998 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-c2v9x-config-2pzlq"] Oct 04 03:21:08 crc kubenswrapper[4770]: W1004 03:21:08.899745 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc2cdaf1_afea_47f5_ab60_a236cfb20e7c.slice/crio-cb4bceb4af9fbca64de9c34bb60c0310062ab48adf8483081c0b7f6acb86afed WatchSource:0}: Error finding container cb4bceb4af9fbca64de9c34bb60c0310062ab48adf8483081c0b7f6acb86afed: Status 404 returned error can't find the container with id cb4bceb4af9fbca64de9c34bb60c0310062ab48adf8483081c0b7f6acb86afed Oct 04 03:21:09 crc kubenswrapper[4770]: I1004 03:21:09.715117 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31252013-1c9c-4586-a45b-e693a1149181" path="/var/lib/kubelet/pods/31252013-1c9c-4586-a45b-e693a1149181/volumes" Oct 04 03:21:09 crc kubenswrapper[4770]: I1004 03:21:09.751507 4770 generic.go:334] "Generic (PLEG): container finished" podID="fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" containerID="f46efdc0729af256ce5b27b4d468a815cdb31c7d04265184e0b0b624114f04fa" exitCode=0 Oct 04 03:21:09 crc kubenswrapper[4770]: I1004 03:21:09.751616 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x-config-2pzlq" event={"ID":"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c","Type":"ContainerDied","Data":"f46efdc0729af256ce5b27b4d468a815cdb31c7d04265184e0b0b624114f04fa"} Oct 04 03:21:09 crc kubenswrapper[4770]: I1004 03:21:09.751674 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x-config-2pzlq" event={"ID":"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c","Type":"ContainerStarted","Data":"cb4bceb4af9fbca64de9c34bb60c0310062ab48adf8483081c0b7f6acb86afed"} Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.205246 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.508810 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-slblm"] Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.510227 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-slblm" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.528454 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-slblm"] Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.596367 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf522\" (UniqueName: \"kubernetes.io/projected/8063ce88-486b-413b-b13f-ac3c855fce17-kube-api-access-cf522\") pod \"cinder-db-create-slblm\" (UID: \"8063ce88-486b-413b-b13f-ac3c855fce17\") " pod="openstack/cinder-db-create-slblm" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.607807 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-v872z"] Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.608881 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v872z" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.619914 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v872z"] Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.697926 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r52s\" (UniqueName: \"kubernetes.io/projected/a0443425-c134-49da-963c-79c4188e9780-kube-api-access-4r52s\") pod \"barbican-db-create-v872z\" (UID: \"a0443425-c134-49da-963c-79c4188e9780\") " pod="openstack/barbican-db-create-v872z" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.698071 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf522\" (UniqueName: \"kubernetes.io/projected/8063ce88-486b-413b-b13f-ac3c855fce17-kube-api-access-cf522\") pod \"cinder-db-create-slblm\" (UID: \"8063ce88-486b-413b-b13f-ac3c855fce17\") " pod="openstack/cinder-db-create-slblm" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.719945 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf522\" (UniqueName: \"kubernetes.io/projected/8063ce88-486b-413b-b13f-ac3c855fce17-kube-api-access-cf522\") pod \"cinder-db-create-slblm\" (UID: \"8063ce88-486b-413b-b13f-ac3c855fce17\") " pod="openstack/cinder-db-create-slblm" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.799392 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r52s\" (UniqueName: \"kubernetes.io/projected/a0443425-c134-49da-963c-79c4188e9780-kube-api-access-4r52s\") pod \"barbican-db-create-v872z\" (UID: \"a0443425-c134-49da-963c-79c4188e9780\") " pod="openstack/barbican-db-create-v872z" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.805892 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-wnwk5"] Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.807047 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wnwk5" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.820958 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wnwk5"] Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.831610 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-slblm" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.839415 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r52s\" (UniqueName: \"kubernetes.io/projected/a0443425-c134-49da-963c-79c4188e9780-kube-api-access-4r52s\") pod \"barbican-db-create-v872z\" (UID: \"a0443425-c134-49da-963c-79c4188e9780\") " pod="openstack/barbican-db-create-v872z" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.880232 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.901453 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4btz\" (UniqueName: \"kubernetes.io/projected/666786ce-354d-4d76-bf21-80d3245b7bec-kube-api-access-l4btz\") pod \"neutron-db-create-wnwk5\" (UID: \"666786ce-354d-4d76-bf21-80d3245b7bec\") " pod="openstack/neutron-db-create-wnwk5" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.926813 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v872z" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.955114 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-29vw4"] Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.956076 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.959358 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.959437 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wdvhv" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.959851 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.959974 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 03:21:14 crc kubenswrapper[4770]: I1004 03:21:14.976487 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-29vw4"] Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.003394 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4btz\" (UniqueName: \"kubernetes.io/projected/666786ce-354d-4d76-bf21-80d3245b7bec-kube-api-access-l4btz\") pod \"neutron-db-create-wnwk5\" (UID: \"666786ce-354d-4d76-bf21-80d3245b7bec\") " pod="openstack/neutron-db-create-wnwk5" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.025614 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4btz\" (UniqueName: \"kubernetes.io/projected/666786ce-354d-4d76-bf21-80d3245b7bec-kube-api-access-l4btz\") pod \"neutron-db-create-wnwk5\" (UID: \"666786ce-354d-4d76-bf21-80d3245b7bec\") " pod="openstack/neutron-db-create-wnwk5" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.104720 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s287f\" (UniqueName: \"kubernetes.io/projected/d4952962-0913-4c58-8647-9ff6823a2553-kube-api-access-s287f\") pod \"keystone-db-sync-29vw4\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.104796 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-config-data\") pod \"keystone-db-sync-29vw4\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.104873 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-combined-ca-bundle\") pod \"keystone-db-sync-29vw4\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.130491 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wnwk5" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.206749 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-config-data\") pod \"keystone-db-sync-29vw4\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.206835 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-combined-ca-bundle\") pod \"keystone-db-sync-29vw4\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.206896 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s287f\" (UniqueName: \"kubernetes.io/projected/d4952962-0913-4c58-8647-9ff6823a2553-kube-api-access-s287f\") pod \"keystone-db-sync-29vw4\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.215945 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-config-data\") pod \"keystone-db-sync-29vw4\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.223377 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-combined-ca-bundle\") pod \"keystone-db-sync-29vw4\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.226704 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s287f\" (UniqueName: \"kubernetes.io/projected/d4952962-0913-4c58-8647-9ff6823a2553-kube-api-access-s287f\") pod \"keystone-db-sync-29vw4\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:15 crc kubenswrapper[4770]: I1004 03:21:15.278871 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:16 crc kubenswrapper[4770]: I1004 03:21:16.286301 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:16 crc kubenswrapper[4770]: I1004 03:21:16.344321 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-d5np4"] Oct 04 03:21:16 crc kubenswrapper[4770]: I1004 03:21:16.344648 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" podUID="e31826ad-abb4-4996-ab4e-49185b5f88dd" containerName="dnsmasq-dns" containerID="cri-o://caffb128695c98b74c1d1bbd2233e7ab75314b870bc7b1697cfec5bc0cbfcbbd" gracePeriod=10 Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.813304 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.850930 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run\") pod \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.851086 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run-ovn\") pod \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.851229 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf4r8\" (UniqueName: \"kubernetes.io/projected/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-kube-api-access-pf4r8\") pod \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.851297 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-log-ovn\") pod \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.851358 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-scripts\") pod \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.851432 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-additional-scripts\") pod \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\" (UID: \"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c\") " Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.853777 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" (UID: "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.853867 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run" (OuterVolumeSpecName: "var-run") pod "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" (UID: "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.853895 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" (UID: "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.863252 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-kube-api-access-pf4r8" (OuterVolumeSpecName: "kube-api-access-pf4r8") pod "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" (UID: "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c"). InnerVolumeSpecName "kube-api-access-pf4r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.863355 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" (UID: "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.864620 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-scripts" (OuterVolumeSpecName: "scripts") pod "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" (UID: "fc2cdaf1-afea-47f5-ab60-a236cfb20e7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.871561 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x-config-2pzlq" event={"ID":"fc2cdaf1-afea-47f5-ab60-a236cfb20e7c","Type":"ContainerDied","Data":"cb4bceb4af9fbca64de9c34bb60c0310062ab48adf8483081c0b7f6acb86afed"} Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.871612 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb4bceb4af9fbca64de9c34bb60c0310062ab48adf8483081c0b7f6acb86afed" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.871711 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x-config-2pzlq" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.896916 4770 generic.go:334] "Generic (PLEG): container finished" podID="e31826ad-abb4-4996-ab4e-49185b5f88dd" containerID="caffb128695c98b74c1d1bbd2233e7ab75314b870bc7b1697cfec5bc0cbfcbbd" exitCode=0 Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.896966 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" event={"ID":"e31826ad-abb4-4996-ab4e-49185b5f88dd","Type":"ContainerDied","Data":"caffb128695c98b74c1d1bbd2233e7ab75314b870bc7b1697cfec5bc0cbfcbbd"} Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.962871 4770 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.969229 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.969261 4770 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.969280 4770 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.969296 4770 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:17 crc kubenswrapper[4770]: I1004 03:21:17.969309 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf4r8\" (UniqueName: \"kubernetes.io/projected/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c-kube-api-access-pf4r8\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.099697 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.172297 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-sb\") pod \"e31826ad-abb4-4996-ab4e-49185b5f88dd\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.172463 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-dns-svc\") pod \"e31826ad-abb4-4996-ab4e-49185b5f88dd\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.172515 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2rxh\" (UniqueName: \"kubernetes.io/projected/e31826ad-abb4-4996-ab4e-49185b5f88dd-kube-api-access-d2rxh\") pod \"e31826ad-abb4-4996-ab4e-49185b5f88dd\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.172550 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-config\") pod \"e31826ad-abb4-4996-ab4e-49185b5f88dd\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.172639 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-nb\") pod \"e31826ad-abb4-4996-ab4e-49185b5f88dd\" (UID: \"e31826ad-abb4-4996-ab4e-49185b5f88dd\") " Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.181064 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31826ad-abb4-4996-ab4e-49185b5f88dd-kube-api-access-d2rxh" (OuterVolumeSpecName: "kube-api-access-d2rxh") pod "e31826ad-abb4-4996-ab4e-49185b5f88dd" (UID: "e31826ad-abb4-4996-ab4e-49185b5f88dd"). InnerVolumeSpecName "kube-api-access-d2rxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.221065 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e31826ad-abb4-4996-ab4e-49185b5f88dd" (UID: "e31826ad-abb4-4996-ab4e-49185b5f88dd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.234272 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-config" (OuterVolumeSpecName: "config") pod "e31826ad-abb4-4996-ab4e-49185b5f88dd" (UID: "e31826ad-abb4-4996-ab4e-49185b5f88dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.239422 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e31826ad-abb4-4996-ab4e-49185b5f88dd" (UID: "e31826ad-abb4-4996-ab4e-49185b5f88dd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.250161 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e31826ad-abb4-4996-ab4e-49185b5f88dd" (UID: "e31826ad-abb4-4996-ab4e-49185b5f88dd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.275025 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.280252 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.280274 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.280284 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2rxh\" (UniqueName: \"kubernetes.io/projected/e31826ad-abb4-4996-ab4e-49185b5f88dd-kube-api-access-d2rxh\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.280297 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e31826ad-abb4-4996-ab4e-49185b5f88dd-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.460408 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v872z"] Oct 04 03:21:18 crc kubenswrapper[4770]: W1004 03:21:18.473083 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0443425_c134_49da_963c_79c4188e9780.slice/crio-eb03946fe078a6b652e96f0d16b06d6161170903fd272b66f1629f72d39b4c7d WatchSource:0}: Error finding container eb03946fe078a6b652e96f0d16b06d6161170903fd272b66f1629f72d39b4c7d: Status 404 returned error can't find the container with id eb03946fe078a6b652e96f0d16b06d6161170903fd272b66f1629f72d39b4c7d Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.576769 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-29vw4"] Oct 04 03:21:18 crc kubenswrapper[4770]: W1004 03:21:18.591595 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod666786ce_354d_4d76_bf21_80d3245b7bec.slice/crio-b1e966b13fbd37ab8bf579efb9973f89bd673fbe68a52b20316ba7e21cc3fcdb WatchSource:0}: Error finding container b1e966b13fbd37ab8bf579efb9973f89bd673fbe68a52b20316ba7e21cc3fcdb: Status 404 returned error can't find the container with id b1e966b13fbd37ab8bf579efb9973f89bd673fbe68a52b20316ba7e21cc3fcdb Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.593485 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wnwk5"] Oct 04 03:21:18 crc kubenswrapper[4770]: W1004 03:21:18.597292 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8063ce88_486b_413b_b13f_ac3c855fce17.slice/crio-7a5849784fb246fe1e79a54559d4598b9557bc9c58b2795413fed0249817b800 WatchSource:0}: Error finding container 7a5849784fb246fe1e79a54559d4598b9557bc9c58b2795413fed0249817b800: Status 404 returned error can't find the container with id 7a5849784fb246fe1e79a54559d4598b9557bc9c58b2795413fed0249817b800 Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.606344 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-slblm"] Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.890927 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-c2v9x-config-2pzlq"] Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.902024 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-c2v9x-config-2pzlq"] Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.914124 4770 generic.go:334] "Generic (PLEG): container finished" podID="a0443425-c134-49da-963c-79c4188e9780" containerID="970002f335d604db0f240fc84212bcb06e3f17bed2f69add957b8bf4503ff2be" exitCode=0 Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.914233 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v872z" event={"ID":"a0443425-c134-49da-963c-79c4188e9780","Type":"ContainerDied","Data":"970002f335d604db0f240fc84212bcb06e3f17bed2f69add957b8bf4503ff2be"} Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.914334 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v872z" event={"ID":"a0443425-c134-49da-963c-79c4188e9780","Type":"ContainerStarted","Data":"eb03946fe078a6b652e96f0d16b06d6161170903fd272b66f1629f72d39b4c7d"} Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.916788 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-slblm" event={"ID":"8063ce88-486b-413b-b13f-ac3c855fce17","Type":"ContainerStarted","Data":"7a5849784fb246fe1e79a54559d4598b9557bc9c58b2795413fed0249817b800"} Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.922236 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-29vw4" event={"ID":"d4952962-0913-4c58-8647-9ff6823a2553","Type":"ContainerStarted","Data":"8e7bc26f1dc601776d8c4ae52ee3e8fc37ebcd9f04b04ede1c4c8423aa8f54af"} Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.942919 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" event={"ID":"e31826ad-abb4-4996-ab4e-49185b5f88dd","Type":"ContainerDied","Data":"0abe1b17d87d19f76498cd5e9d09167c838803a9303bf34ba77feefd175b2041"} Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.942975 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665d79f5c9-d5np4" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.942981 4770 scope.go:117] "RemoveContainer" containerID="caffb128695c98b74c1d1bbd2233e7ab75314b870bc7b1697cfec5bc0cbfcbbd" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.947487 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ql7kt" event={"ID":"76095968-a5a2-4a39-bf86-5faafb9d76d3","Type":"ContainerStarted","Data":"48b401920dbda52bc727cb8f3066f846fcf5688b17b9700513de47eeb995f7fb"} Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.950752 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wnwk5" event={"ID":"666786ce-354d-4d76-bf21-80d3245b7bec","Type":"ContainerStarted","Data":"b1e966b13fbd37ab8bf579efb9973f89bd673fbe68a52b20316ba7e21cc3fcdb"} Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.968782 4770 scope.go:117] "RemoveContainer" containerID="eddfca7d848951d50e3fe2267ab70c8c527f62e474d01a2a10a096c3f9ffa905" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.974830 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-ql7kt" podStartSLOduration=2.694288639 podStartE2EDuration="15.973056659s" podCreationTimestamp="2025-10-04 03:21:03 +0000 UTC" firstStartedPulling="2025-10-04 03:21:04.725200834 +0000 UTC m=+1076.017210576" lastFinishedPulling="2025-10-04 03:21:18.003968894 +0000 UTC m=+1089.295978596" observedRunningTime="2025-10-04 03:21:18.964424164 +0000 UTC m=+1090.256433886" watchObservedRunningTime="2025-10-04 03:21:18.973056659 +0000 UTC m=+1090.265066381" Oct 04 03:21:18 crc kubenswrapper[4770]: I1004 03:21:18.998656 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-wnwk5" podStartSLOduration=4.998639079 podStartE2EDuration="4.998639079s" podCreationTimestamp="2025-10-04 03:21:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:21:18.984336955 +0000 UTC m=+1090.276346677" watchObservedRunningTime="2025-10-04 03:21:18.998639079 +0000 UTC m=+1090.290648801" Oct 04 03:21:19 crc kubenswrapper[4770]: I1004 03:21:19.006805 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-d5np4"] Oct 04 03:21:19 crc kubenswrapper[4770]: I1004 03:21:19.013672 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-665d79f5c9-d5np4"] Oct 04 03:21:19 crc kubenswrapper[4770]: I1004 03:21:19.687241 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31826ad-abb4-4996-ab4e-49185b5f88dd" path="/var/lib/kubelet/pods/e31826ad-abb4-4996-ab4e-49185b5f88dd/volumes" Oct 04 03:21:19 crc kubenswrapper[4770]: I1004 03:21:19.688030 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" path="/var/lib/kubelet/pods/fc2cdaf1-afea-47f5-ab60-a236cfb20e7c/volumes" Oct 04 03:21:19 crc kubenswrapper[4770]: I1004 03:21:19.962251 4770 generic.go:334] "Generic (PLEG): container finished" podID="8063ce88-486b-413b-b13f-ac3c855fce17" containerID="3fa4c602bfba3b93aa690acefd5cac4104a43a1685f6fc8bf4e1a15387371281" exitCode=0 Oct 04 03:21:19 crc kubenswrapper[4770]: I1004 03:21:19.962354 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-slblm" event={"ID":"8063ce88-486b-413b-b13f-ac3c855fce17","Type":"ContainerDied","Data":"3fa4c602bfba3b93aa690acefd5cac4104a43a1685f6fc8bf4e1a15387371281"} Oct 04 03:21:19 crc kubenswrapper[4770]: I1004 03:21:19.964871 4770 generic.go:334] "Generic (PLEG): container finished" podID="666786ce-354d-4d76-bf21-80d3245b7bec" containerID="35253e9b3bb51aa3385e4b6549e518d5109f6cc1177b4048be361a4ce14ca70b" exitCode=0 Oct 04 03:21:19 crc kubenswrapper[4770]: I1004 03:21:19.965128 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wnwk5" event={"ID":"666786ce-354d-4d76-bf21-80d3245b7bec","Type":"ContainerDied","Data":"35253e9b3bb51aa3385e4b6549e518d5109f6cc1177b4048be361a4ce14ca70b"} Oct 04 03:21:20 crc kubenswrapper[4770]: I1004 03:21:20.267695 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v872z" Oct 04 03:21:20 crc kubenswrapper[4770]: I1004 03:21:20.318127 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r52s\" (UniqueName: \"kubernetes.io/projected/a0443425-c134-49da-963c-79c4188e9780-kube-api-access-4r52s\") pod \"a0443425-c134-49da-963c-79c4188e9780\" (UID: \"a0443425-c134-49da-963c-79c4188e9780\") " Oct 04 03:21:20 crc kubenswrapper[4770]: I1004 03:21:20.324641 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0443425-c134-49da-963c-79c4188e9780-kube-api-access-4r52s" (OuterVolumeSpecName: "kube-api-access-4r52s") pod "a0443425-c134-49da-963c-79c4188e9780" (UID: "a0443425-c134-49da-963c-79c4188e9780"). InnerVolumeSpecName "kube-api-access-4r52s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:20 crc kubenswrapper[4770]: I1004 03:21:20.419601 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r52s\" (UniqueName: \"kubernetes.io/projected/a0443425-c134-49da-963c-79c4188e9780-kube-api-access-4r52s\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:20 crc kubenswrapper[4770]: I1004 03:21:20.975443 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v872z" Oct 04 03:21:20 crc kubenswrapper[4770]: I1004 03:21:20.985320 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v872z" event={"ID":"a0443425-c134-49da-963c-79c4188e9780","Type":"ContainerDied","Data":"eb03946fe078a6b652e96f0d16b06d6161170903fd272b66f1629f72d39b4c7d"} Oct 04 03:21:20 crc kubenswrapper[4770]: I1004 03:21:20.985359 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb03946fe078a6b652e96f0d16b06d6161170903fd272b66f1629f72d39b4c7d" Oct 04 03:21:23 crc kubenswrapper[4770]: I1004 03:21:23.271184 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-slblm" Oct 04 03:21:23 crc kubenswrapper[4770]: I1004 03:21:23.299484 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wnwk5" Oct 04 03:21:23 crc kubenswrapper[4770]: I1004 03:21:23.371326 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf522\" (UniqueName: \"kubernetes.io/projected/8063ce88-486b-413b-b13f-ac3c855fce17-kube-api-access-cf522\") pod \"8063ce88-486b-413b-b13f-ac3c855fce17\" (UID: \"8063ce88-486b-413b-b13f-ac3c855fce17\") " Oct 04 03:21:23 crc kubenswrapper[4770]: I1004 03:21:23.371609 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4btz\" (UniqueName: \"kubernetes.io/projected/666786ce-354d-4d76-bf21-80d3245b7bec-kube-api-access-l4btz\") pod \"666786ce-354d-4d76-bf21-80d3245b7bec\" (UID: \"666786ce-354d-4d76-bf21-80d3245b7bec\") " Oct 04 03:21:23 crc kubenswrapper[4770]: I1004 03:21:23.375296 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/666786ce-354d-4d76-bf21-80d3245b7bec-kube-api-access-l4btz" (OuterVolumeSpecName: "kube-api-access-l4btz") pod "666786ce-354d-4d76-bf21-80d3245b7bec" (UID: "666786ce-354d-4d76-bf21-80d3245b7bec"). InnerVolumeSpecName "kube-api-access-l4btz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:23 crc kubenswrapper[4770]: I1004 03:21:23.379734 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8063ce88-486b-413b-b13f-ac3c855fce17-kube-api-access-cf522" (OuterVolumeSpecName: "kube-api-access-cf522") pod "8063ce88-486b-413b-b13f-ac3c855fce17" (UID: "8063ce88-486b-413b-b13f-ac3c855fce17"). InnerVolumeSpecName "kube-api-access-cf522". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:23 crc kubenswrapper[4770]: I1004 03:21:23.473906 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4btz\" (UniqueName: \"kubernetes.io/projected/666786ce-354d-4d76-bf21-80d3245b7bec-kube-api-access-l4btz\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:23 crc kubenswrapper[4770]: I1004 03:21:23.473958 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf522\" (UniqueName: \"kubernetes.io/projected/8063ce88-486b-413b-b13f-ac3c855fce17-kube-api-access-cf522\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.005827 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-slblm" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.005854 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-slblm" event={"ID":"8063ce88-486b-413b-b13f-ac3c855fce17","Type":"ContainerDied","Data":"7a5849784fb246fe1e79a54559d4598b9557bc9c58b2795413fed0249817b800"} Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.005924 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a5849784fb246fe1e79a54559d4598b9557bc9c58b2795413fed0249817b800" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.008395 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wnwk5" event={"ID":"666786ce-354d-4d76-bf21-80d3245b7bec","Type":"ContainerDied","Data":"b1e966b13fbd37ab8bf579efb9973f89bd673fbe68a52b20316ba7e21cc3fcdb"} Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.008430 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1e966b13fbd37ab8bf579efb9973f89bd673fbe68a52b20316ba7e21cc3fcdb" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.008502 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wnwk5" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.011935 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-29vw4" event={"ID":"d4952962-0913-4c58-8647-9ff6823a2553","Type":"ContainerStarted","Data":"5adbac3cda39aabfa24a6562bafcc4eaada11c462c1b8f6a9852131751821625"} Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.308551 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-29vw4" podStartSLOduration=5.789227942 podStartE2EDuration="10.308533721s" podCreationTimestamp="2025-10-04 03:21:14 +0000 UTC" firstStartedPulling="2025-10-04 03:21:18.612553303 +0000 UTC m=+1089.904563015" lastFinishedPulling="2025-10-04 03:21:23.131859062 +0000 UTC m=+1094.423868794" observedRunningTime="2025-10-04 03:21:24.040895095 +0000 UTC m=+1095.332904817" watchObservedRunningTime="2025-10-04 03:21:24.308533721 +0000 UTC m=+1095.600543443" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.635700 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec07-account-create-zfvpg"] Oct 04 03:21:24 crc kubenswrapper[4770]: E1004 03:21:24.636040 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31826ad-abb4-4996-ab4e-49185b5f88dd" containerName="dnsmasq-dns" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636052 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31826ad-abb4-4996-ab4e-49185b5f88dd" containerName="dnsmasq-dns" Oct 04 03:21:24 crc kubenswrapper[4770]: E1004 03:21:24.636066 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8063ce88-486b-413b-b13f-ac3c855fce17" containerName="mariadb-database-create" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636072 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8063ce88-486b-413b-b13f-ac3c855fce17" containerName="mariadb-database-create" Oct 04 03:21:24 crc kubenswrapper[4770]: E1004 03:21:24.636089 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666786ce-354d-4d76-bf21-80d3245b7bec" containerName="mariadb-database-create" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636096 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="666786ce-354d-4d76-bf21-80d3245b7bec" containerName="mariadb-database-create" Oct 04 03:21:24 crc kubenswrapper[4770]: E1004 03:21:24.636114 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31826ad-abb4-4996-ab4e-49185b5f88dd" containerName="init" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636120 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31826ad-abb4-4996-ab4e-49185b5f88dd" containerName="init" Oct 04 03:21:24 crc kubenswrapper[4770]: E1004 03:21:24.636133 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" containerName="ovn-config" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636138 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" containerName="ovn-config" Oct 04 03:21:24 crc kubenswrapper[4770]: E1004 03:21:24.636149 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0443425-c134-49da-963c-79c4188e9780" containerName="mariadb-database-create" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636154 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0443425-c134-49da-963c-79c4188e9780" containerName="mariadb-database-create" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636303 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31826ad-abb4-4996-ab4e-49185b5f88dd" containerName="dnsmasq-dns" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636311 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc2cdaf1-afea-47f5-ab60-a236cfb20e7c" containerName="ovn-config" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636321 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0443425-c134-49da-963c-79c4188e9780" containerName="mariadb-database-create" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636331 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="666786ce-354d-4d76-bf21-80d3245b7bec" containerName="mariadb-database-create" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636342 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8063ce88-486b-413b-b13f-ac3c855fce17" containerName="mariadb-database-create" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.636896 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec07-account-create-zfvpg" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.651617 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec07-account-create-zfvpg"] Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.655178 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.693255 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql44d\" (UniqueName: \"kubernetes.io/projected/8acc0024-f6e4-44a4-a76d-007c7ff3ecd8-kube-api-access-ql44d\") pod \"cinder-ec07-account-create-zfvpg\" (UID: \"8acc0024-f6e4-44a4-a76d-007c7ff3ecd8\") " pod="openstack/cinder-ec07-account-create-zfvpg" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.795041 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql44d\" (UniqueName: \"kubernetes.io/projected/8acc0024-f6e4-44a4-a76d-007c7ff3ecd8-kube-api-access-ql44d\") pod \"cinder-ec07-account-create-zfvpg\" (UID: \"8acc0024-f6e4-44a4-a76d-007c7ff3ecd8\") " pod="openstack/cinder-ec07-account-create-zfvpg" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.822416 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql44d\" (UniqueName: \"kubernetes.io/projected/8acc0024-f6e4-44a4-a76d-007c7ff3ecd8-kube-api-access-ql44d\") pod \"cinder-ec07-account-create-zfvpg\" (UID: \"8acc0024-f6e4-44a4-a76d-007c7ff3ecd8\") " pod="openstack/cinder-ec07-account-create-zfvpg" Oct 04 03:21:24 crc kubenswrapper[4770]: I1004 03:21:24.954241 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec07-account-create-zfvpg" Oct 04 03:21:25 crc kubenswrapper[4770]: I1004 03:21:25.426353 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec07-account-create-zfvpg"] Oct 04 03:21:25 crc kubenswrapper[4770]: W1004 03:21:25.452408 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8acc0024_f6e4_44a4_a76d_007c7ff3ecd8.slice/crio-769e2e60866440b53602eb1c6b49e226fcb272702edc27de4279d504b3862507 WatchSource:0}: Error finding container 769e2e60866440b53602eb1c6b49e226fcb272702edc27de4279d504b3862507: Status 404 returned error can't find the container with id 769e2e60866440b53602eb1c6b49e226fcb272702edc27de4279d504b3862507 Oct 04 03:21:26 crc kubenswrapper[4770]: I1004 03:21:26.032266 4770 generic.go:334] "Generic (PLEG): container finished" podID="76095968-a5a2-4a39-bf86-5faafb9d76d3" containerID="48b401920dbda52bc727cb8f3066f846fcf5688b17b9700513de47eeb995f7fb" exitCode=0 Oct 04 03:21:26 crc kubenswrapper[4770]: I1004 03:21:26.032392 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ql7kt" event={"ID":"76095968-a5a2-4a39-bf86-5faafb9d76d3","Type":"ContainerDied","Data":"48b401920dbda52bc727cb8f3066f846fcf5688b17b9700513de47eeb995f7fb"} Oct 04 03:21:26 crc kubenswrapper[4770]: I1004 03:21:26.034097 4770 generic.go:334] "Generic (PLEG): container finished" podID="8acc0024-f6e4-44a4-a76d-007c7ff3ecd8" containerID="1c08379885b40ec9939052a4eeecf97debed59c25f447022d90963addf4dcf17" exitCode=0 Oct 04 03:21:26 crc kubenswrapper[4770]: I1004 03:21:26.034153 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec07-account-create-zfvpg" event={"ID":"8acc0024-f6e4-44a4-a76d-007c7ff3ecd8","Type":"ContainerDied","Data":"1c08379885b40ec9939052a4eeecf97debed59c25f447022d90963addf4dcf17"} Oct 04 03:21:26 crc kubenswrapper[4770]: I1004 03:21:26.034178 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec07-account-create-zfvpg" event={"ID":"8acc0024-f6e4-44a4-a76d-007c7ff3ecd8","Type":"ContainerStarted","Data":"769e2e60866440b53602eb1c6b49e226fcb272702edc27de4279d504b3862507"} Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.048936 4770 generic.go:334] "Generic (PLEG): container finished" podID="d4952962-0913-4c58-8647-9ff6823a2553" containerID="5adbac3cda39aabfa24a6562bafcc4eaada11c462c1b8f6a9852131751821625" exitCode=0 Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.049069 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-29vw4" event={"ID":"d4952962-0913-4c58-8647-9ff6823a2553","Type":"ContainerDied","Data":"5adbac3cda39aabfa24a6562bafcc4eaada11c462c1b8f6a9852131751821625"} Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.449961 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec07-account-create-zfvpg" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.541841 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql44d\" (UniqueName: \"kubernetes.io/projected/8acc0024-f6e4-44a4-a76d-007c7ff3ecd8-kube-api-access-ql44d\") pod \"8acc0024-f6e4-44a4-a76d-007c7ff3ecd8\" (UID: \"8acc0024-f6e4-44a4-a76d-007c7ff3ecd8\") " Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.549524 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8acc0024-f6e4-44a4-a76d-007c7ff3ecd8-kube-api-access-ql44d" (OuterVolumeSpecName: "kube-api-access-ql44d") pod "8acc0024-f6e4-44a4-a76d-007c7ff3ecd8" (UID: "8acc0024-f6e4-44a4-a76d-007c7ff3ecd8"). InnerVolumeSpecName "kube-api-access-ql44d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.608055 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.643518 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-config-data\") pod \"76095968-a5a2-4a39-bf86-5faafb9d76d3\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.643657 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2cpr\" (UniqueName: \"kubernetes.io/projected/76095968-a5a2-4a39-bf86-5faafb9d76d3-kube-api-access-h2cpr\") pod \"76095968-a5a2-4a39-bf86-5faafb9d76d3\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.643697 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-combined-ca-bundle\") pod \"76095968-a5a2-4a39-bf86-5faafb9d76d3\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.643783 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-db-sync-config-data\") pod \"76095968-a5a2-4a39-bf86-5faafb9d76d3\" (UID: \"76095968-a5a2-4a39-bf86-5faafb9d76d3\") " Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.645323 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql44d\" (UniqueName: \"kubernetes.io/projected/8acc0024-f6e4-44a4-a76d-007c7ff3ecd8-kube-api-access-ql44d\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.649671 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "76095968-a5a2-4a39-bf86-5faafb9d76d3" (UID: "76095968-a5a2-4a39-bf86-5faafb9d76d3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.650213 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76095968-a5a2-4a39-bf86-5faafb9d76d3-kube-api-access-h2cpr" (OuterVolumeSpecName: "kube-api-access-h2cpr") pod "76095968-a5a2-4a39-bf86-5faafb9d76d3" (UID: "76095968-a5a2-4a39-bf86-5faafb9d76d3"). InnerVolumeSpecName "kube-api-access-h2cpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.669349 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76095968-a5a2-4a39-bf86-5faafb9d76d3" (UID: "76095968-a5a2-4a39-bf86-5faafb9d76d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.689800 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-config-data" (OuterVolumeSpecName: "config-data") pod "76095968-a5a2-4a39-bf86-5faafb9d76d3" (UID: "76095968-a5a2-4a39-bf86-5faafb9d76d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.748161 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.748200 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2cpr\" (UniqueName: \"kubernetes.io/projected/76095968-a5a2-4a39-bf86-5faafb9d76d3-kube-api-access-h2cpr\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.748212 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:27 crc kubenswrapper[4770]: I1004 03:21:27.748224 4770 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76095968-a5a2-4a39-bf86-5faafb9d76d3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.058674 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec07-account-create-zfvpg" event={"ID":"8acc0024-f6e4-44a4-a76d-007c7ff3ecd8","Type":"ContainerDied","Data":"769e2e60866440b53602eb1c6b49e226fcb272702edc27de4279d504b3862507"} Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.058720 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="769e2e60866440b53602eb1c6b49e226fcb272702edc27de4279d504b3862507" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.058690 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec07-account-create-zfvpg" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.060944 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ql7kt" event={"ID":"76095968-a5a2-4a39-bf86-5faafb9d76d3","Type":"ContainerDied","Data":"e3abcaa6dd73d67eb01e1e9b20df49fdbaa75c4b93f855918976362a19c494d4"} Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.060967 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3abcaa6dd73d67eb01e1e9b20df49fdbaa75c4b93f855918976362a19c494d4" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.060980 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ql7kt" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.491793 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58486c5949-trj2z"] Oct 04 03:21:28 crc kubenswrapper[4770]: E1004 03:21:28.492891 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76095968-a5a2-4a39-bf86-5faafb9d76d3" containerName="glance-db-sync" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.492917 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="76095968-a5a2-4a39-bf86-5faafb9d76d3" containerName="glance-db-sync" Oct 04 03:21:28 crc kubenswrapper[4770]: E1004 03:21:28.492949 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8acc0024-f6e4-44a4-a76d-007c7ff3ecd8" containerName="mariadb-account-create" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.492958 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8acc0024-f6e4-44a4-a76d-007c7ff3ecd8" containerName="mariadb-account-create" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.493197 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="76095968-a5a2-4a39-bf86-5faafb9d76d3" containerName="glance-db-sync" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.493222 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8acc0024-f6e4-44a4-a76d-007c7ff3ecd8" containerName="mariadb-account-create" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.494661 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.542138 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58486c5949-trj2z"] Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.673099 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-config\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.673149 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-swift-storage-0\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.673171 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-sb\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.673332 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-svc\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.673379 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-nb\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.673404 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc7w7\" (UniqueName: \"kubernetes.io/projected/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-kube-api-access-cc7w7\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.736629 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.775770 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-svc\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.775841 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-nb\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.775882 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc7w7\" (UniqueName: \"kubernetes.io/projected/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-kube-api-access-cc7w7\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.776283 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-config\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.776313 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-swift-storage-0\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.776729 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-svc\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.777126 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-config\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.777301 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-sb\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.777446 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-nb\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.777884 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-swift-storage-0\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.778636 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-sb\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.802582 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc7w7\" (UniqueName: \"kubernetes.io/projected/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-kube-api-access-cc7w7\") pod \"dnsmasq-dns-58486c5949-trj2z\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.823563 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.878691 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-combined-ca-bundle\") pod \"d4952962-0913-4c58-8647-9ff6823a2553\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.879251 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s287f\" (UniqueName: \"kubernetes.io/projected/d4952962-0913-4c58-8647-9ff6823a2553-kube-api-access-s287f\") pod \"d4952962-0913-4c58-8647-9ff6823a2553\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.879308 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-config-data\") pod \"d4952962-0913-4c58-8647-9ff6823a2553\" (UID: \"d4952962-0913-4c58-8647-9ff6823a2553\") " Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.884469 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4952962-0913-4c58-8647-9ff6823a2553-kube-api-access-s287f" (OuterVolumeSpecName: "kube-api-access-s287f") pod "d4952962-0913-4c58-8647-9ff6823a2553" (UID: "d4952962-0913-4c58-8647-9ff6823a2553"). InnerVolumeSpecName "kube-api-access-s287f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.907742 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4952962-0913-4c58-8647-9ff6823a2553" (UID: "d4952962-0913-4c58-8647-9ff6823a2553"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.977130 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-config-data" (OuterVolumeSpecName: "config-data") pod "d4952962-0913-4c58-8647-9ff6823a2553" (UID: "d4952962-0913-4c58-8647-9ff6823a2553"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.982157 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s287f\" (UniqueName: \"kubernetes.io/projected/d4952962-0913-4c58-8647-9ff6823a2553-kube-api-access-s287f\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.982182 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:28 crc kubenswrapper[4770]: I1004 03:21:28.982193 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4952962-0913-4c58-8647-9ff6823a2553-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.082944 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-29vw4" event={"ID":"d4952962-0913-4c58-8647-9ff6823a2553","Type":"ContainerDied","Data":"8e7bc26f1dc601776d8c4ae52ee3e8fc37ebcd9f04b04ede1c4c8423aa8f54af"} Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.083119 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e7bc26f1dc601776d8c4ae52ee3e8fc37ebcd9f04b04ede1c4c8423aa8f54af" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.083124 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-29vw4" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.322083 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58486c5949-trj2z"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.328773 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-thblw"] Oct 04 03:21:29 crc kubenswrapper[4770]: E1004 03:21:29.329265 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4952962-0913-4c58-8647-9ff6823a2553" containerName="keystone-db-sync" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.329289 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4952962-0913-4c58-8647-9ff6823a2553" containerName="keystone-db-sync" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.329513 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4952962-0913-4c58-8647-9ff6823a2553" containerName="keystone-db-sync" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.331445 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.339483 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.339759 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wdvhv" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.340066 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.340322 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.360974 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-thblw"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.402775 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58486c5949-trj2z"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.441971 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c88d9674c-hrsnk"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.443499 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.489327 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c88d9674c-hrsnk"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.491886 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lxrw\" (UniqueName: \"kubernetes.io/projected/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-kube-api-access-4lxrw\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.491946 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-config-data\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.491997 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-combined-ca-bundle\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.492054 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-fernet-keys\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.492101 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-scripts\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.492130 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-credential-keys\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.594902 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-combined-ca-bundle\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595492 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-sb\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595520 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmv9k\" (UniqueName: \"kubernetes.io/projected/2d5a0829-d679-488e-8ef6-93770e70076d-kube-api-access-nmv9k\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595563 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-fernet-keys\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595606 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-svc\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595632 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-scripts\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595674 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-credential-keys\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595697 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-nb\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595722 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lxrw\" (UniqueName: \"kubernetes.io/projected/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-kube-api-access-4lxrw\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595742 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-config\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595766 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-config-data\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.595794 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-swift-storage-0\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.615125 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.617780 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-fernet-keys\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.619137 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-credential-keys\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.629835 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lxrw\" (UniqueName: \"kubernetes.io/projected/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-kube-api-access-4lxrw\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.631800 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-combined-ca-bundle\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.637127 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-scripts\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.649823 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-config-data\") pod \"keystone-bootstrap-thblw\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.678643 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.705430 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.712207 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.738166 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-config\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.738326 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-swift-storage-0\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.738480 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-sb\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.738536 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmv9k\" (UniqueName: \"kubernetes.io/projected/2d5a0829-d679-488e-8ef6-93770e70076d-kube-api-access-nmv9k\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.738743 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-svc\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.738890 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-nb\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.739904 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-nb\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.740439 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-swift-storage-0\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.743737 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-sb\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.744422 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-svc\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.747077 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-config\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.751428 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.765132 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c88d9674c-hrsnk"] Oct 04 03:21:29 crc kubenswrapper[4770]: E1004 03:21:29.766147 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-nmv9k], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" podUID="2d5a0829-d679-488e-8ef6-93770e70076d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.768194 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmv9k\" (UniqueName: \"kubernetes.io/projected/2d5a0829-d679-488e-8ef6-93770e70076d-kube-api-access-nmv9k\") pod \"dnsmasq-dns-7c88d9674c-hrsnk\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.779236 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6886d"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.780508 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.782110 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.782811 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9vxhs" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.783292 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.783354 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.788085 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6795bd975-h4h4h"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.793351 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.811142 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-h4h4h"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.820137 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6886d"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.842320 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-scripts\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.842520 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-logs\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.842633 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-log-httpd\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.842708 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-config-data\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.842777 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.842851 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-swift-storage-0\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.842915 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-svc\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.842980 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-combined-ca-bundle\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843057 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tqql\" (UniqueName: \"kubernetes.io/projected/13c7863f-832e-4a86-8ce6-f57c92e834c6-kube-api-access-7tqql\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843121 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-config\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843190 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr6hw\" (UniqueName: \"kubernetes.io/projected/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-kube-api-access-xr6hw\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843257 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-sb\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843328 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-nb\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843400 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pj5t\" (UniqueName: \"kubernetes.io/projected/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-kube-api-access-4pj5t\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843478 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-config-data\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843644 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843716 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-run-httpd\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.843791 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-scripts\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.855834 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-hkqts"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.857532 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.860514 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.860684 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.860841 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gpljl" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.868810 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hkqts"] Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.945987 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-config-data\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946383 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946412 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-combined-ca-bundle\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946445 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-swift-storage-0\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946472 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-svc\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946492 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-combined-ca-bundle\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946517 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tqql\" (UniqueName: \"kubernetes.io/projected/13c7863f-832e-4a86-8ce6-f57c92e834c6-kube-api-access-7tqql\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946536 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-scripts\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946555 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-config\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946582 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr6hw\" (UniqueName: \"kubernetes.io/projected/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-kube-api-access-xr6hw\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-sb\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946629 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-nb\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946654 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pj5t\" (UniqueName: \"kubernetes.io/projected/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-kube-api-access-4pj5t\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946671 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dck9k\" (UniqueName: \"kubernetes.io/projected/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-kube-api-access-dck9k\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946700 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-config-data\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946724 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-config-data\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946742 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-db-sync-config-data\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946765 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946784 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-run-httpd\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946807 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-scripts\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946829 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-scripts\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946845 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-logs\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946872 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-etc-machine-id\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.946889 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-log-httpd\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.947631 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-log-httpd\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.949653 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-nb\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.952864 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-run-httpd\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.952885 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-svc\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.953482 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-config\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.953918 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-logs\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.954456 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-sb\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.954865 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.955417 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-config-data\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.957486 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.961186 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-config-data\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.961233 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-scripts\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.961771 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-scripts\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.962452 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-swift-storage-0\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.966995 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pj5t\" (UniqueName: \"kubernetes.io/projected/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-kube-api-access-4pj5t\") pod \"ceilometer-0\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " pod="openstack/ceilometer-0" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.970837 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr6hw\" (UniqueName: \"kubernetes.io/projected/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-kube-api-access-xr6hw\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.971089 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tqql\" (UniqueName: \"kubernetes.io/projected/13c7863f-832e-4a86-8ce6-f57c92e834c6-kube-api-access-7tqql\") pod \"dnsmasq-dns-6795bd975-h4h4h\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:29 crc kubenswrapper[4770]: I1004 03:21:29.971986 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-combined-ca-bundle\") pod \"placement-db-sync-6886d\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " pod="openstack/placement-db-sync-6886d" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.047627 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-combined-ca-bundle\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.047705 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-scripts\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.047756 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dck9k\" (UniqueName: \"kubernetes.io/projected/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-kube-api-access-dck9k\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.047791 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-config-data\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.047812 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-db-sync-config-data\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.047857 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-etc-machine-id\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.047942 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-etc-machine-id\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.054269 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-scripts\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.055099 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-db-sync-config-data\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.056805 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-combined-ca-bundle\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.057066 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-config-data\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.062531 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.072817 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dck9k\" (UniqueName: \"kubernetes.io/projected/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-kube-api-access-dck9k\") pod \"cinder-db-sync-hkqts\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.106132 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-thblw"] Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.108581 4770 generic.go:334] "Generic (PLEG): container finished" podID="2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" containerID="c72064d13e32842d54f892f85c13b9765080991f9cf8a0be2a16a38edc18b0c9" exitCode=0 Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.108675 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.109365 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58486c5949-trj2z" event={"ID":"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f","Type":"ContainerDied","Data":"c72064d13e32842d54f892f85c13b9765080991f9cf8a0be2a16a38edc18b0c9"} Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.109391 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58486c5949-trj2z" event={"ID":"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f","Type":"ContainerStarted","Data":"06f2dcbaa56206764d951dd016e25342181b01a6669ebc0db31a5179a4f9cacb"} Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.126897 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6886d" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.141142 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.148458 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.180397 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hkqts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.258462 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-sb\") pod \"2d5a0829-d679-488e-8ef6-93770e70076d\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.258916 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-swift-storage-0\") pod \"2d5a0829-d679-488e-8ef6-93770e70076d\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.258982 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-nb\") pod \"2d5a0829-d679-488e-8ef6-93770e70076d\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.259038 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-config\") pod \"2d5a0829-d679-488e-8ef6-93770e70076d\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.259113 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmv9k\" (UniqueName: \"kubernetes.io/projected/2d5a0829-d679-488e-8ef6-93770e70076d-kube-api-access-nmv9k\") pod \"2d5a0829-d679-488e-8ef6-93770e70076d\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.259198 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-svc\") pod \"2d5a0829-d679-488e-8ef6-93770e70076d\" (UID: \"2d5a0829-d679-488e-8ef6-93770e70076d\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.259604 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2d5a0829-d679-488e-8ef6-93770e70076d" (UID: "2d5a0829-d679-488e-8ef6-93770e70076d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.260113 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2d5a0829-d679-488e-8ef6-93770e70076d" (UID: "2d5a0829-d679-488e-8ef6-93770e70076d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.260466 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2d5a0829-d679-488e-8ef6-93770e70076d" (UID: "2d5a0829-d679-488e-8ef6-93770e70076d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.263342 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-config" (OuterVolumeSpecName: "config") pod "2d5a0829-d679-488e-8ef6-93770e70076d" (UID: "2d5a0829-d679-488e-8ef6-93770e70076d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.267154 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d5a0829-d679-488e-8ef6-93770e70076d-kube-api-access-nmv9k" (OuterVolumeSpecName: "kube-api-access-nmv9k") pod "2d5a0829-d679-488e-8ef6-93770e70076d" (UID: "2d5a0829-d679-488e-8ef6-93770e70076d"). InnerVolumeSpecName "kube-api-access-nmv9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.268233 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d5a0829-d679-488e-8ef6-93770e70076d" (UID: "2d5a0829-d679-488e-8ef6-93770e70076d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.360950 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.360980 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.360992 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.361013 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.361022 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmv9k\" (UniqueName: \"kubernetes.io/projected/2d5a0829-d679-488e-8ef6-93770e70076d-kube-api-access-nmv9k\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.361031 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d5a0829-d679-488e-8ef6-93770e70076d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.453546 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.454758 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.468812 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.468992 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mc7pb" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.469126 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.491620 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.538368 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.539868 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.544983 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.546699 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.564245 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-logs\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.564310 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-config-data\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.564363 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.564588 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.564608 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.564646 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-755x7\" (UniqueName: \"kubernetes.io/projected/b848edbd-09d2-4807-90b0-31cd8bcf9816-kube-api-access-755x7\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.564677 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-scripts\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.638807 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.665994 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666041 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666061 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-755x7\" (UniqueName: \"kubernetes.io/projected/b848edbd-09d2-4807-90b0-31cd8bcf9816-kube-api-access-755x7\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666098 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-scripts\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666129 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666151 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666168 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666190 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-logs\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666221 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666247 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-config-data\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666282 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666298 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sjd6\" (UniqueName: \"kubernetes.io/projected/17a5e8dc-16d2-4c72-97de-92200a911cfd-kube-api-access-5sjd6\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666325 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666345 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-logs\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666460 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.666508 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.667278 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-logs\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.675794 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.687650 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-scripts\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.689161 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-config-data\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.712232 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-755x7\" (UniqueName: \"kubernetes.io/projected/b848edbd-09d2-4807-90b0-31cd8bcf9816-kube-api-access-755x7\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.741067 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.768877 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.769246 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sjd6\" (UniqueName: \"kubernetes.io/projected/17a5e8dc-16d2-4c72-97de-92200a911cfd-kube-api-access-5sjd6\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.769369 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-logs\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.769564 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.769644 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.769709 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.769809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.770740 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.771241 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.771410 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-logs\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.778101 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.785371 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.788653 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sjd6\" (UniqueName: \"kubernetes.io/projected/17a5e8dc-16d2-4c72-97de-92200a911cfd-kube-api-access-5sjd6\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.790273 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.800258 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.807761 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.870760 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-config\") pod \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.870899 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-nb\") pod \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.870922 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-sb\") pod \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.870952 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc7w7\" (UniqueName: \"kubernetes.io/projected/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-kube-api-access-cc7w7\") pod \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.870969 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-svc\") pod \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.870984 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-swift-storage-0\") pod \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\" (UID: \"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f\") " Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.877732 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.880550 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-kube-api-access-cc7w7" (OuterVolumeSpecName: "kube-api-access-cc7w7") pod "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" (UID: "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f"). InnerVolumeSpecName "kube-api-access-cc7w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.895590 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.905531 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" (UID: "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.909886 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" (UID: "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.920869 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" (UID: "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.930713 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-config" (OuterVolumeSpecName: "config") pod "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" (UID: "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.933359 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" (UID: "2bff14ff-ad5e-4454-8a62-fbd8c7d1264f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.972788 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.973117 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.973129 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.973138 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc7w7\" (UniqueName: \"kubernetes.io/projected/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-kube-api-access-cc7w7\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.973147 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.973156 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.979281 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-h4h4h"] Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.984230 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hkqts"] Oct 04 03:21:30 crc kubenswrapper[4770]: I1004 03:21:30.992819 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6886d"] Oct 04 03:21:31 crc kubenswrapper[4770]: W1004 03:21:31.045671 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a8836e2_0ea8_43cb_ba3f_535b106c4f3a.slice/crio-548c9e38e8dd3d3ead2d773d631f6995aed01505bf068f91b52fc28517698384 WatchSource:0}: Error finding container 548c9e38e8dd3d3ead2d773d631f6995aed01505bf068f91b52fc28517698384: Status 404 returned error can't find the container with id 548c9e38e8dd3d3ead2d773d631f6995aed01505bf068f91b52fc28517698384 Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.125682 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-thblw" event={"ID":"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae","Type":"ContainerStarted","Data":"edfad2a88ea7cdcc0ab69cc5032405bd7fcbabf51fba14a38595d0a638d62106"} Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.125730 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-thblw" event={"ID":"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae","Type":"ContainerStarted","Data":"aabae12de20526a32e1b111a65bc726c0f3de37b3ccfef95e3dd729e99eb950a"} Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.128818 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerStarted","Data":"e8de094b904b664a55471c569a961f61cd5c3bd12b3e7a876d52f3b0987800fc"} Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.136559 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hkqts" event={"ID":"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a","Type":"ContainerStarted","Data":"548c9e38e8dd3d3ead2d773d631f6995aed01505bf068f91b52fc28517698384"} Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.137550 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" event={"ID":"13c7863f-832e-4a86-8ce6-f57c92e834c6","Type":"ContainerStarted","Data":"151b88d45affba4f10727851e4b55864ae77a058e266690913507173a6bf5625"} Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.138496 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6886d" event={"ID":"ec2a2801-f563-4bbf-b191-7e4dffbe89a2","Type":"ContainerStarted","Data":"b8680169542392a169d98c36fd81c8b4071374105a56218cb2029e71339b8154"} Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.139522 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c88d9674c-hrsnk" Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.149183 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58486c5949-trj2z" event={"ID":"2bff14ff-ad5e-4454-8a62-fbd8c7d1264f","Type":"ContainerDied","Data":"06f2dcbaa56206764d951dd016e25342181b01a6669ebc0db31a5179a4f9cacb"} Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.149272 4770 scope.go:117] "RemoveContainer" containerID="c72064d13e32842d54f892f85c13b9765080991f9cf8a0be2a16a38edc18b0c9" Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.149208 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58486c5949-trj2z" Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.153025 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-thblw" podStartSLOduration=2.152995248 podStartE2EDuration="2.152995248s" podCreationTimestamp="2025-10-04 03:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:21:31.149471605 +0000 UTC m=+1102.441481317" watchObservedRunningTime="2025-10-04 03:21:31.152995248 +0000 UTC m=+1102.445004960" Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.206399 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c88d9674c-hrsnk"] Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.220321 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c88d9674c-hrsnk"] Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.264949 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58486c5949-trj2z"] Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.282548 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58486c5949-trj2z"] Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.483399 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.684950 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" path="/var/lib/kubelet/pods/2bff14ff-ad5e-4454-8a62-fbd8c7d1264f/volumes" Oct 04 03:21:31 crc kubenswrapper[4770]: I1004 03:21:31.685759 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d5a0829-d679-488e-8ef6-93770e70076d" path="/var/lib/kubelet/pods/2d5a0829-d679-488e-8ef6-93770e70076d/volumes" Oct 04 03:21:32 crc kubenswrapper[4770]: I1004 03:21:32.155249 4770 generic.go:334] "Generic (PLEG): container finished" podID="13c7863f-832e-4a86-8ce6-f57c92e834c6" containerID="aff796e9bd542c911e26288c4a0a0e2f7cee9640b0e3b0afe91a5d63dc41800a" exitCode=0 Oct 04 03:21:32 crc kubenswrapper[4770]: I1004 03:21:32.155453 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" event={"ID":"13c7863f-832e-4a86-8ce6-f57c92e834c6","Type":"ContainerDied","Data":"aff796e9bd542c911e26288c4a0a0e2f7cee9640b0e3b0afe91a5d63dc41800a"} Oct 04 03:21:32 crc kubenswrapper[4770]: I1004 03:21:32.165442 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b848edbd-09d2-4807-90b0-31cd8bcf9816","Type":"ContainerStarted","Data":"feff1bb5a8fc20a29fadd1ff0bcce4101ec4c377f1d68f2d233252f6a90af3be"} Oct 04 03:21:32 crc kubenswrapper[4770]: I1004 03:21:32.186832 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.177540 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" event={"ID":"13c7863f-832e-4a86-8ce6-f57c92e834c6","Type":"ContainerStarted","Data":"0259add0c4189d05b42efd063fd05d0d7b7cec99654b19c19844b1ff27eb64c4"} Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.180267 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.180304 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17a5e8dc-16d2-4c72-97de-92200a911cfd","Type":"ContainerStarted","Data":"6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54"} Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.180328 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17a5e8dc-16d2-4c72-97de-92200a911cfd","Type":"ContainerStarted","Data":"5f784af38b2f0c0f256d3614a47680819e6bdad1dcafc8c23a7feda648c95b63"} Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.190367 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b848edbd-09d2-4807-90b0-31cd8bcf9816","Type":"ContainerStarted","Data":"475aa92bed7d3e48b456882c7936b86ecaee6d04b9299c91ed9778d6780c69f1"} Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.190416 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b848edbd-09d2-4807-90b0-31cd8bcf9816","Type":"ContainerStarted","Data":"e317ffd093915d7d79e5e83daaab8921d0c0be691df765e1c791cb4978907b04"} Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.203305 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" podStartSLOduration=4.203287478 podStartE2EDuration="4.203287478s" podCreationTimestamp="2025-10-04 03:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:21:33.200458404 +0000 UTC m=+1104.492468106" watchObservedRunningTime="2025-10-04 03:21:33.203287478 +0000 UTC m=+1104.495297190" Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.236302 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.236268671 podStartE2EDuration="4.236268671s" podCreationTimestamp="2025-10-04 03:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:21:33.229382241 +0000 UTC m=+1104.521391953" watchObservedRunningTime="2025-10-04 03:21:33.236268671 +0000 UTC m=+1104.528278383" Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.616725 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.650696 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:21:33 crc kubenswrapper[4770]: I1004 03:21:33.703402 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.206931 4770 generic.go:334] "Generic (PLEG): container finished" podID="a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" containerID="edfad2a88ea7cdcc0ab69cc5032405bd7fcbabf51fba14a38595d0a638d62106" exitCode=0 Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.207806 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-thblw" event={"ID":"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae","Type":"ContainerDied","Data":"edfad2a88ea7cdcc0ab69cc5032405bd7fcbabf51fba14a38595d0a638d62106"} Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.569142 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5365-account-create-47djg"] Oct 04 03:21:34 crc kubenswrapper[4770]: E1004 03:21:34.569563 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" containerName="init" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.569576 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" containerName="init" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.569753 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bff14ff-ad5e-4454-8a62-fbd8c7d1264f" containerName="init" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.570310 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5365-account-create-47djg" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.574841 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.577407 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5365-account-create-47djg"] Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.645472 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc855\" (UniqueName: \"kubernetes.io/projected/483a5eae-40e5-4e11-82e8-24198ec8a733-kube-api-access-bc855\") pod \"barbican-5365-account-create-47djg\" (UID: \"483a5eae-40e5-4e11-82e8-24198ec8a733\") " pod="openstack/barbican-5365-account-create-47djg" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.747480 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc855\" (UniqueName: \"kubernetes.io/projected/483a5eae-40e5-4e11-82e8-24198ec8a733-kube-api-access-bc855\") pod \"barbican-5365-account-create-47djg\" (UID: \"483a5eae-40e5-4e11-82e8-24198ec8a733\") " pod="openstack/barbican-5365-account-create-47djg" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.764667 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b0c9-account-create-dx2vq"] Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.765847 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b0c9-account-create-dx2vq" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.768170 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.782725 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc855\" (UniqueName: \"kubernetes.io/projected/483a5eae-40e5-4e11-82e8-24198ec8a733-kube-api-access-bc855\") pod \"barbican-5365-account-create-47djg\" (UID: \"483a5eae-40e5-4e11-82e8-24198ec8a733\") " pod="openstack/barbican-5365-account-create-47djg" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.792309 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b0c9-account-create-dx2vq"] Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.849591 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frm4p\" (UniqueName: \"kubernetes.io/projected/9c51889c-6112-4c86-93f1-a69876016f4f-kube-api-access-frm4p\") pod \"neutron-b0c9-account-create-dx2vq\" (UID: \"9c51889c-6112-4c86-93f1-a69876016f4f\") " pod="openstack/neutron-b0c9-account-create-dx2vq" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.898500 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5365-account-create-47djg" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.954335 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frm4p\" (UniqueName: \"kubernetes.io/projected/9c51889c-6112-4c86-93f1-a69876016f4f-kube-api-access-frm4p\") pod \"neutron-b0c9-account-create-dx2vq\" (UID: \"9c51889c-6112-4c86-93f1-a69876016f4f\") " pod="openstack/neutron-b0c9-account-create-dx2vq" Oct 04 03:21:34 crc kubenswrapper[4770]: I1004 03:21:34.990582 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frm4p\" (UniqueName: \"kubernetes.io/projected/9c51889c-6112-4c86-93f1-a69876016f4f-kube-api-access-frm4p\") pod \"neutron-b0c9-account-create-dx2vq\" (UID: \"9c51889c-6112-4c86-93f1-a69876016f4f\") " pod="openstack/neutron-b0c9-account-create-dx2vq" Oct 04 03:21:35 crc kubenswrapper[4770]: I1004 03:21:35.143471 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b0c9-account-create-dx2vq" Oct 04 03:21:35 crc kubenswrapper[4770]: I1004 03:21:35.234638 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerName="glance-log" containerID="cri-o://e317ffd093915d7d79e5e83daaab8921d0c0be691df765e1c791cb4978907b04" gracePeriod=30 Oct 04 03:21:35 crc kubenswrapper[4770]: I1004 03:21:35.234807 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerName="glance-httpd" containerID="cri-o://475aa92bed7d3e48b456882c7936b86ecaee6d04b9299c91ed9778d6780c69f1" gracePeriod=30 Oct 04 03:21:36 crc kubenswrapper[4770]: I1004 03:21:36.245469 4770 generic.go:334] "Generic (PLEG): container finished" podID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerID="475aa92bed7d3e48b456882c7936b86ecaee6d04b9299c91ed9778d6780c69f1" exitCode=0 Oct 04 03:21:36 crc kubenswrapper[4770]: I1004 03:21:36.245957 4770 generic.go:334] "Generic (PLEG): container finished" podID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerID="e317ffd093915d7d79e5e83daaab8921d0c0be691df765e1c791cb4978907b04" exitCode=143 Oct 04 03:21:36 crc kubenswrapper[4770]: I1004 03:21:36.245570 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b848edbd-09d2-4807-90b0-31cd8bcf9816","Type":"ContainerDied","Data":"475aa92bed7d3e48b456882c7936b86ecaee6d04b9299c91ed9778d6780c69f1"} Oct 04 03:21:36 crc kubenswrapper[4770]: I1004 03:21:36.246034 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b848edbd-09d2-4807-90b0-31cd8bcf9816","Type":"ContainerDied","Data":"e317ffd093915d7d79e5e83daaab8921d0c0be691df765e1c791cb4978907b04"} Oct 04 03:21:37 crc kubenswrapper[4770]: I1004 03:21:37.810994 4770 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod31252013-1c9c-4586-a45b-e693a1149181"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod31252013-1c9c-4586-a45b-e693a1149181] : Timed out while waiting for systemd to remove kubepods-besteffort-pod31252013_1c9c_4586_a45b_e693a1149181.slice" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.066643 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.222730 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-config-data\") pod \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.223367 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-scripts\") pod \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.223435 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-credential-keys\") pod \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.223522 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-fernet-keys\") pod \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.223575 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lxrw\" (UniqueName: \"kubernetes.io/projected/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-kube-api-access-4lxrw\") pod \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.223602 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-combined-ca-bundle\") pod \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\" (UID: \"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.230380 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" (UID: "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.230859 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-scripts" (OuterVolumeSpecName: "scripts") pod "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" (UID: "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.233197 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-kube-api-access-4lxrw" (OuterVolumeSpecName: "kube-api-access-4lxrw") pod "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" (UID: "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae"). InnerVolumeSpecName "kube-api-access-4lxrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.240819 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" (UID: "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.257766 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" (UID: "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.295563 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-config-data" (OuterVolumeSpecName: "config-data") pod "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" (UID: "a9ae84bc-8bf7-4518-bb2d-a834edbe69ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.297695 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-thblw" event={"ID":"a9ae84bc-8bf7-4518-bb2d-a834edbe69ae","Type":"ContainerDied","Data":"aabae12de20526a32e1b111a65bc726c0f3de37b3ccfef95e3dd729e99eb950a"} Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.297737 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aabae12de20526a32e1b111a65bc726c0f3de37b3ccfef95e3dd729e99eb950a" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.297865 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-thblw" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.325537 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.325564 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lxrw\" (UniqueName: \"kubernetes.io/projected/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-kube-api-access-4lxrw\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.325575 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.325585 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.325594 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.325602 4770 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.719407 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.834591 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-logs\") pod \"b848edbd-09d2-4807-90b0-31cd8bcf9816\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.834870 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-scripts\") pod \"b848edbd-09d2-4807-90b0-31cd8bcf9816\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.834981 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-config-data\") pod \"b848edbd-09d2-4807-90b0-31cd8bcf9816\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.835088 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"b848edbd-09d2-4807-90b0-31cd8bcf9816\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.835170 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-755x7\" (UniqueName: \"kubernetes.io/projected/b848edbd-09d2-4807-90b0-31cd8bcf9816-kube-api-access-755x7\") pod \"b848edbd-09d2-4807-90b0-31cd8bcf9816\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.835277 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-combined-ca-bundle\") pod \"b848edbd-09d2-4807-90b0-31cd8bcf9816\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.835379 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-httpd-run\") pod \"b848edbd-09d2-4807-90b0-31cd8bcf9816\" (UID: \"b848edbd-09d2-4807-90b0-31cd8bcf9816\") " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.835301 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-logs" (OuterVolumeSpecName: "logs") pod "b848edbd-09d2-4807-90b0-31cd8bcf9816" (UID: "b848edbd-09d2-4807-90b0-31cd8bcf9816"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.835980 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b848edbd-09d2-4807-90b0-31cd8bcf9816" (UID: "b848edbd-09d2-4807-90b0-31cd8bcf9816"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.841755 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "b848edbd-09d2-4807-90b0-31cd8bcf9816" (UID: "b848edbd-09d2-4807-90b0-31cd8bcf9816"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.842266 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-scripts" (OuterVolumeSpecName: "scripts") pod "b848edbd-09d2-4807-90b0-31cd8bcf9816" (UID: "b848edbd-09d2-4807-90b0-31cd8bcf9816"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.844535 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b848edbd-09d2-4807-90b0-31cd8bcf9816-kube-api-access-755x7" (OuterVolumeSpecName: "kube-api-access-755x7") pod "b848edbd-09d2-4807-90b0-31cd8bcf9816" (UID: "b848edbd-09d2-4807-90b0-31cd8bcf9816"). InnerVolumeSpecName "kube-api-access-755x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.872746 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b848edbd-09d2-4807-90b0-31cd8bcf9816" (UID: "b848edbd-09d2-4807-90b0-31cd8bcf9816"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.884396 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-config-data" (OuterVolumeSpecName: "config-data") pod "b848edbd-09d2-4807-90b0-31cd8bcf9816" (UID: "b848edbd-09d2-4807-90b0-31cd8bcf9816"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.936579 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.936614 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.936624 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b848edbd-09d2-4807-90b0-31cd8bcf9816-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.936632 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.936639 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b848edbd-09d2-4807-90b0-31cd8bcf9816-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.936671 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.936704 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-755x7\" (UniqueName: \"kubernetes.io/projected/b848edbd-09d2-4807-90b0-31cd8bcf9816-kube-api-access-755x7\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:39 crc kubenswrapper[4770]: I1004 03:21:39.957515 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.037925 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.150916 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.201957 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-9f4f4"] Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.202201 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" podUID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerName="dnsmasq-dns" containerID="cri-o://95c4469d744bdeafad6ffb6c0b084279944cb12985e2d21884cf1967cf113de9" gracePeriod=10 Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.230113 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-thblw"] Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.248811 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-thblw"] Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.307390 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nq9d8"] Oct 04 03:21:40 crc kubenswrapper[4770]: E1004 03:21:40.310453 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerName="glance-httpd" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.310480 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerName="glance-httpd" Oct 04 03:21:40 crc kubenswrapper[4770]: E1004 03:21:40.310607 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" containerName="keystone-bootstrap" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.310616 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" containerName="keystone-bootstrap" Oct 04 03:21:40 crc kubenswrapper[4770]: E1004 03:21:40.310652 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerName="glance-log" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.310694 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerName="glance-log" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.311250 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerName="glance-log" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.311286 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" containerName="keystone-bootstrap" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.311309 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b848edbd-09d2-4807-90b0-31cd8bcf9816" containerName="glance-httpd" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.312029 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.316402 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.316875 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.317049 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.317704 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wdvhv" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.320509 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b848edbd-09d2-4807-90b0-31cd8bcf9816","Type":"ContainerDied","Data":"feff1bb5a8fc20a29fadd1ff0bcce4101ec4c377f1d68f2d233252f6a90af3be"} Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.320590 4770 scope.go:117] "RemoveContainer" containerID="475aa92bed7d3e48b456882c7936b86ecaee6d04b9299c91ed9778d6780c69f1" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.320880 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.375297 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nq9d8"] Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.427556 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.441999 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.449804 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd2nz\" (UniqueName: \"kubernetes.io/projected/c9dfa4b6-223e-447d-b62c-1f772c82802d-kube-api-access-jd2nz\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.449893 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-credential-keys\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.449968 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-combined-ca-bundle\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.450116 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-fernet-keys\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.450187 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-scripts\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.450412 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-config-data\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.451965 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.456458 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.460040 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.460443 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.466051 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.551785 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-scripts\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.551853 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-combined-ca-bundle\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.551926 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-fernet-keys\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.551968 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-scripts\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552036 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552137 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-config-data\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552164 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552203 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd2nz\" (UniqueName: \"kubernetes.io/projected/c9dfa4b6-223e-447d-b62c-1f772c82802d-kube-api-access-jd2nz\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552288 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-config-data\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552312 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-logs\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552342 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552383 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552413 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhsl6\" (UniqueName: \"kubernetes.io/projected/d4377592-806c-4c05-9527-fe09c7802b73-kube-api-access-zhsl6\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.552483 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-credential-keys\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.556976 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-credential-keys\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.557298 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-fernet-keys\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.568600 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-config-data\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.569819 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-scripts\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.570372 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-combined-ca-bundle\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.575705 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd2nz\" (UniqueName: \"kubernetes.io/projected/c9dfa4b6-223e-447d-b62c-1f772c82802d-kube-api-access-jd2nz\") pod \"keystone-bootstrap-nq9d8\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.641196 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.653922 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-config-data\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.653974 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-logs\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.654025 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.654064 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.654093 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhsl6\" (UniqueName: \"kubernetes.io/projected/d4377592-806c-4c05-9527-fe09c7802b73-kube-api-access-zhsl6\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.654152 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-scripts\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.654197 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.654256 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.654836 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.658442 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.658584 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-logs\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.659666 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-config-data\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.666392 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.677541 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-scripts\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.685973 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.688846 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhsl6\" (UniqueName: \"kubernetes.io/projected/d4377592-806c-4c05-9527-fe09c7802b73-kube-api-access-zhsl6\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.690432 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " pod="openstack/glance-default-external-api-0" Oct 04 03:21:40 crc kubenswrapper[4770]: I1004 03:21:40.841351 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:21:41 crc kubenswrapper[4770]: I1004 03:21:41.284968 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" podUID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: connect: connection refused" Oct 04 03:21:41 crc kubenswrapper[4770]: I1004 03:21:41.335354 4770 generic.go:334] "Generic (PLEG): container finished" podID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerID="95c4469d744bdeafad6ffb6c0b084279944cb12985e2d21884cf1967cf113de9" exitCode=0 Oct 04 03:21:41 crc kubenswrapper[4770]: I1004 03:21:41.335432 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" event={"ID":"03359d36-a52b-4969-b1c6-38eaf57d89fa","Type":"ContainerDied","Data":"95c4469d744bdeafad6ffb6c0b084279944cb12985e2d21884cf1967cf113de9"} Oct 04 03:21:41 crc kubenswrapper[4770]: I1004 03:21:41.686031 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ae84bc-8bf7-4518-bb2d-a834edbe69ae" path="/var/lib/kubelet/pods/a9ae84bc-8bf7-4518-bb2d-a834edbe69ae/volumes" Oct 04 03:21:41 crc kubenswrapper[4770]: I1004 03:21:41.687749 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b848edbd-09d2-4807-90b0-31cd8bcf9816" path="/var/lib/kubelet/pods/b848edbd-09d2-4807-90b0-31cd8bcf9816/volumes" Oct 04 03:21:46 crc kubenswrapper[4770]: I1004 03:21:46.283906 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" podUID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: connect: connection refused" Oct 04 03:21:48 crc kubenswrapper[4770]: I1004 03:21:48.222238 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b0c9-account-create-dx2vq"] Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.077064 4770 scope.go:117] "RemoveContainer" containerID="e317ffd093915d7d79e5e83daaab8921d0c0be691df765e1c791cb4978907b04" Oct 04 03:21:49 crc kubenswrapper[4770]: E1004 03:21:49.080808 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213" Oct 04 03:21:49 crc kubenswrapper[4770]: E1004 03:21:49.081121 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dck9k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-hkqts_openstack(3a8836e2-0ea8-43cb-ba3f-535b106c4f3a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:21:49 crc kubenswrapper[4770]: E1004 03:21:49.082455 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-hkqts" podUID="3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" Oct 04 03:21:49 crc kubenswrapper[4770]: W1004 03:21:49.145778 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c51889c_6112_4c86_93f1_a69876016f4f.slice/crio-2850a2d6bebae745e198bea1b48f42b4aa3a069ad8c8562782d94fe1c5c485a2 WatchSource:0}: Error finding container 2850a2d6bebae745e198bea1b48f42b4aa3a069ad8c8562782d94fe1c5c485a2: Status 404 returned error can't find the container with id 2850a2d6bebae745e198bea1b48f42b4aa3a069ad8c8562782d94fe1c5c485a2 Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.222175 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.355722 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4jfg\" (UniqueName: \"kubernetes.io/projected/03359d36-a52b-4969-b1c6-38eaf57d89fa-kube-api-access-k4jfg\") pod \"03359d36-a52b-4969-b1c6-38eaf57d89fa\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.355823 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-config\") pod \"03359d36-a52b-4969-b1c6-38eaf57d89fa\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.355845 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-nb\") pod \"03359d36-a52b-4969-b1c6-38eaf57d89fa\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.355902 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-svc\") pod \"03359d36-a52b-4969-b1c6-38eaf57d89fa\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.355935 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-swift-storage-0\") pod \"03359d36-a52b-4969-b1c6-38eaf57d89fa\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.355958 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-sb\") pod \"03359d36-a52b-4969-b1c6-38eaf57d89fa\" (UID: \"03359d36-a52b-4969-b1c6-38eaf57d89fa\") " Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.362068 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03359d36-a52b-4969-b1c6-38eaf57d89fa-kube-api-access-k4jfg" (OuterVolumeSpecName: "kube-api-access-k4jfg") pod "03359d36-a52b-4969-b1c6-38eaf57d89fa" (UID: "03359d36-a52b-4969-b1c6-38eaf57d89fa"). InnerVolumeSpecName "kube-api-access-k4jfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.457526 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4jfg\" (UniqueName: \"kubernetes.io/projected/03359d36-a52b-4969-b1c6-38eaf57d89fa-kube-api-access-k4jfg\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.468037 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.469033 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77594f5649-9f4f4" event={"ID":"03359d36-a52b-4969-b1c6-38eaf57d89fa","Type":"ContainerDied","Data":"fb7b8cc9baf0021ac1683dd10800180be95ceb9590fbde17ce68b45b48cd6e83"} Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.469117 4770 scope.go:117] "RemoveContainer" containerID="95c4469d744bdeafad6ffb6c0b084279944cb12985e2d21884cf1967cf113de9" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.478246 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b0c9-account-create-dx2vq" event={"ID":"9c51889c-6112-4c86-93f1-a69876016f4f","Type":"ContainerStarted","Data":"2850a2d6bebae745e198bea1b48f42b4aa3a069ad8c8562782d94fe1c5c485a2"} Oct 04 03:21:49 crc kubenswrapper[4770]: E1004 03:21:49.480649 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213\\\"\"" pod="openstack/cinder-db-sync-hkqts" podUID="3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.514365 4770 scope.go:117] "RemoveContainer" containerID="cdfe922c4750454f542d0e693ca4bec380e1bd0655e34c774bacd5179bdeae43" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.550626 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "03359d36-a52b-4969-b1c6-38eaf57d89fa" (UID: "03359d36-a52b-4969-b1c6-38eaf57d89fa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.561628 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-config" (OuterVolumeSpecName: "config") pod "03359d36-a52b-4969-b1c6-38eaf57d89fa" (UID: "03359d36-a52b-4969-b1c6-38eaf57d89fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.562713 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.562752 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.570060 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "03359d36-a52b-4969-b1c6-38eaf57d89fa" (UID: "03359d36-a52b-4969-b1c6-38eaf57d89fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.570701 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "03359d36-a52b-4969-b1c6-38eaf57d89fa" (UID: "03359d36-a52b-4969-b1c6-38eaf57d89fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.587219 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "03359d36-a52b-4969-b1c6-38eaf57d89fa" (UID: "03359d36-a52b-4969-b1c6-38eaf57d89fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.625264 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5365-account-create-47djg"] Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.664929 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.664964 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.664973 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03359d36-a52b-4969-b1c6-38eaf57d89fa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.779318 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.800345 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nq9d8"] Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.807283 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-9f4f4"] Oct 04 03:21:49 crc kubenswrapper[4770]: I1004 03:21:49.814423 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77594f5649-9f4f4"] Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.514243 4770 generic.go:334] "Generic (PLEG): container finished" podID="483a5eae-40e5-4e11-82e8-24198ec8a733" containerID="edf5ef1a49c54d5bdd50958c0f9f2bd3bad10a29cc67fffa8fd87dd4e3b7e29d" exitCode=0 Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.514795 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5365-account-create-47djg" event={"ID":"483a5eae-40e5-4e11-82e8-24198ec8a733","Type":"ContainerDied","Data":"edf5ef1a49c54d5bdd50958c0f9f2bd3bad10a29cc67fffa8fd87dd4e3b7e29d"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.515522 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5365-account-create-47djg" event={"ID":"483a5eae-40e5-4e11-82e8-24198ec8a733","Type":"ContainerStarted","Data":"ccc6e88fd2c962bea367391e6ce84a3173a80e4602ffe474e9506ad7cae54e8e"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.531409 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17a5e8dc-16d2-4c72-97de-92200a911cfd","Type":"ContainerStarted","Data":"66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.531579 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerName="glance-log" containerID="cri-o://6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54" gracePeriod=30 Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.531637 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerName="glance-httpd" containerID="cri-o://66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392" gracePeriod=30 Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.534298 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d4377592-806c-4c05-9527-fe09c7802b73","Type":"ContainerStarted","Data":"a0f9b2c44a0420584b66e0af2a755d11a35c8d018901f1ba39dec7459258e5a3"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.534335 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d4377592-806c-4c05-9527-fe09c7802b73","Type":"ContainerStarted","Data":"2c2d6b0f430cc36ff0f138569977ab3ba444759f49ad43944b5906f5b23f8350"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.540910 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6886d" event={"ID":"ec2a2801-f563-4bbf-b191-7e4dffbe89a2","Type":"ContainerStarted","Data":"c2e39f05e83ac887adbf3376bdea8800dc2e57fa7efb4f00b67adc9235d689e5"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.542127 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerStarted","Data":"bb46755f85a4e4f356284e969d79b3ca1f637207f6244f543490ca472e39341a"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.546211 4770 generic.go:334] "Generic (PLEG): container finished" podID="9c51889c-6112-4c86-93f1-a69876016f4f" containerID="1d60905ae32f34af588942deabcb007f319002f38b02f98d03e42cf5c89bca33" exitCode=0 Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.546272 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b0c9-account-create-dx2vq" event={"ID":"9c51889c-6112-4c86-93f1-a69876016f4f","Type":"ContainerDied","Data":"1d60905ae32f34af588942deabcb007f319002f38b02f98d03e42cf5c89bca33"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.553953 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nq9d8" event={"ID":"c9dfa4b6-223e-447d-b62c-1f772c82802d","Type":"ContainerStarted","Data":"df651e173b6556cb139982e461ff9c637c9cf1e2459276d87df36652f29f1557"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.554046 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nq9d8" event={"ID":"c9dfa4b6-223e-447d-b62c-1f772c82802d","Type":"ContainerStarted","Data":"2870179140fb8c592ec64ae32730710b3f2e5757309305b952ba6f4faea1cdad"} Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.594435 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=21.594381107 podStartE2EDuration="21.594381107s" podCreationTimestamp="2025-10-04 03:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:21:50.566662262 +0000 UTC m=+1121.858671974" watchObservedRunningTime="2025-10-04 03:21:50.594381107 +0000 UTC m=+1121.886390819" Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.595824 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nq9d8" podStartSLOduration=10.595815085 podStartE2EDuration="10.595815085s" podCreationTimestamp="2025-10-04 03:21:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:21:50.58682386 +0000 UTC m=+1121.878833572" watchObservedRunningTime="2025-10-04 03:21:50.595815085 +0000 UTC m=+1121.887824797" Oct 04 03:21:50 crc kubenswrapper[4770]: I1004 03:21:50.611992 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6886d" podStartSLOduration=4.917187276 podStartE2EDuration="21.611975458s" podCreationTimestamp="2025-10-04 03:21:29 +0000 UTC" firstStartedPulling="2025-10-04 03:21:31.036575232 +0000 UTC m=+1102.328584944" lastFinishedPulling="2025-10-04 03:21:47.731363404 +0000 UTC m=+1119.023373126" observedRunningTime="2025-10-04 03:21:50.61050964 +0000 UTC m=+1121.902519352" watchObservedRunningTime="2025-10-04 03:21:50.611975458 +0000 UTC m=+1121.903985170" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.222786 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.295412 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-config-data\") pod \"17a5e8dc-16d2-4c72-97de-92200a911cfd\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.295483 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sjd6\" (UniqueName: \"kubernetes.io/projected/17a5e8dc-16d2-4c72-97de-92200a911cfd-kube-api-access-5sjd6\") pod \"17a5e8dc-16d2-4c72-97de-92200a911cfd\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.295537 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-scripts\") pod \"17a5e8dc-16d2-4c72-97de-92200a911cfd\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.295570 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-httpd-run\") pod \"17a5e8dc-16d2-4c72-97de-92200a911cfd\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.295696 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"17a5e8dc-16d2-4c72-97de-92200a911cfd\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.295732 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-combined-ca-bundle\") pod \"17a5e8dc-16d2-4c72-97de-92200a911cfd\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.295852 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-logs\") pod \"17a5e8dc-16d2-4c72-97de-92200a911cfd\" (UID: \"17a5e8dc-16d2-4c72-97de-92200a911cfd\") " Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.296328 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-logs" (OuterVolumeSpecName: "logs") pod "17a5e8dc-16d2-4c72-97de-92200a911cfd" (UID: "17a5e8dc-16d2-4c72-97de-92200a911cfd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.296631 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "17a5e8dc-16d2-4c72-97de-92200a911cfd" (UID: "17a5e8dc-16d2-4c72-97de-92200a911cfd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.301206 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "17a5e8dc-16d2-4c72-97de-92200a911cfd" (UID: "17a5e8dc-16d2-4c72-97de-92200a911cfd"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.301386 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17a5e8dc-16d2-4c72-97de-92200a911cfd-kube-api-access-5sjd6" (OuterVolumeSpecName: "kube-api-access-5sjd6") pod "17a5e8dc-16d2-4c72-97de-92200a911cfd" (UID: "17a5e8dc-16d2-4c72-97de-92200a911cfd"). InnerVolumeSpecName "kube-api-access-5sjd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.302415 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-scripts" (OuterVolumeSpecName: "scripts") pod "17a5e8dc-16d2-4c72-97de-92200a911cfd" (UID: "17a5e8dc-16d2-4c72-97de-92200a911cfd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.350304 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-config-data" (OuterVolumeSpecName: "config-data") pod "17a5e8dc-16d2-4c72-97de-92200a911cfd" (UID: "17a5e8dc-16d2-4c72-97de-92200a911cfd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.353115 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17a5e8dc-16d2-4c72-97de-92200a911cfd" (UID: "17a5e8dc-16d2-4c72-97de-92200a911cfd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.398124 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sjd6\" (UniqueName: \"kubernetes.io/projected/17a5e8dc-16d2-4c72-97de-92200a911cfd-kube-api-access-5sjd6\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.398162 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.398177 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.398217 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.398231 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.398246 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17a5e8dc-16d2-4c72-97de-92200a911cfd-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.398259 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17a5e8dc-16d2-4c72-97de-92200a911cfd-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.426315 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.500165 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.564606 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d4377592-806c-4c05-9527-fe09c7802b73","Type":"ContainerStarted","Data":"28014651bae2a240ed227b651ddc14ed3143c2ea8cc47d1b44c7a6b98514b480"} Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.567122 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerStarted","Data":"88fc7604d5d5ddc8c731243d3ea5484edeab92c6728d874de648b554b94a02a8"} Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.569760 4770 generic.go:334] "Generic (PLEG): container finished" podID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerID="66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392" exitCode=0 Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.569791 4770 generic.go:334] "Generic (PLEG): container finished" podID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerID="6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54" exitCode=143 Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.570390 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17a5e8dc-16d2-4c72-97de-92200a911cfd","Type":"ContainerDied","Data":"66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392"} Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.570450 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17a5e8dc-16d2-4c72-97de-92200a911cfd","Type":"ContainerDied","Data":"6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54"} Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.570469 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17a5e8dc-16d2-4c72-97de-92200a911cfd","Type":"ContainerDied","Data":"5f784af38b2f0c0f256d3614a47680819e6bdad1dcafc8c23a7feda648c95b63"} Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.570489 4770 scope.go:117] "RemoveContainer" containerID="66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.570730 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.600477 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.600446919 podStartE2EDuration="11.600446919s" podCreationTimestamp="2025-10-04 03:21:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:21:51.585661762 +0000 UTC m=+1122.877671484" watchObservedRunningTime="2025-10-04 03:21:51.600446919 +0000 UTC m=+1122.892456631" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.640517 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.646381 4770 scope.go:117] "RemoveContainer" containerID="6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.657836 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.668036 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:21:51 crc kubenswrapper[4770]: E1004 03:21:51.668685 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerName="glance-log" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.668702 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerName="glance-log" Oct 04 03:21:51 crc kubenswrapper[4770]: E1004 03:21:51.668714 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerName="glance-httpd" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.668723 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerName="glance-httpd" Oct 04 03:21:51 crc kubenswrapper[4770]: E1004 03:21:51.668749 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerName="dnsmasq-dns" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.668755 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerName="dnsmasq-dns" Oct 04 03:21:51 crc kubenswrapper[4770]: E1004 03:21:51.668779 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerName="init" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.668785 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerName="init" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.668942 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="03359d36-a52b-4969-b1c6-38eaf57d89fa" containerName="dnsmasq-dns" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.668966 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerName="glance-log" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.668976 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="17a5e8dc-16d2-4c72-97de-92200a911cfd" containerName="glance-httpd" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.669940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.674509 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.681309 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.701732 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03359d36-a52b-4969-b1c6-38eaf57d89fa" path="/var/lib/kubelet/pods/03359d36-a52b-4969-b1c6-38eaf57d89fa/volumes" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.704628 4770 scope.go:117] "RemoveContainer" containerID="66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392" Oct 04 03:21:51 crc kubenswrapper[4770]: E1004 03:21:51.709248 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392\": container with ID starting with 66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392 not found: ID does not exist" containerID="66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.709285 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392"} err="failed to get container status \"66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392\": rpc error: code = NotFound desc = could not find container \"66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392\": container with ID starting with 66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392 not found: ID does not exist" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.709312 4770 scope.go:117] "RemoveContainer" containerID="6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54" Oct 04 03:21:51 crc kubenswrapper[4770]: E1004 03:21:51.709927 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54\": container with ID starting with 6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54 not found: ID does not exist" containerID="6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.709969 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54"} err="failed to get container status \"6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54\": rpc error: code = NotFound desc = could not find container \"6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54\": container with ID starting with 6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54 not found: ID does not exist" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.709993 4770 scope.go:117] "RemoveContainer" containerID="66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.710347 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392"} err="failed to get container status \"66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392\": rpc error: code = NotFound desc = could not find container \"66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392\": container with ID starting with 66803ed476d928634077503392b0920d4f0a1a3d0dcfd7aff3b0b591cd149392 not found: ID does not exist" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.710393 4770 scope.go:117] "RemoveContainer" containerID="6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.711082 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17a5e8dc-16d2-4c72-97de-92200a911cfd" path="/var/lib/kubelet/pods/17a5e8dc-16d2-4c72-97de-92200a911cfd/volumes" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.711754 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54"} err="failed to get container status \"6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54\": rpc error: code = NotFound desc = could not find container \"6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54\": container with ID starting with 6a3ec159d62d887bb092919a8958ef7413b99fb380826b4b4bc25e2ba17c7e54 not found: ID does not exist" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.711812 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.805400 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.805454 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.805475 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-logs\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.805508 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.805545 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.805578 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24vzb\" (UniqueName: \"kubernetes.io/projected/94883722-c05e-43f6-ac7d-91ce20de293b-kube-api-access-24vzb\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.805870 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.805906 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.908176 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24vzb\" (UniqueName: \"kubernetes.io/projected/94883722-c05e-43f6-ac7d-91ce20de293b-kube-api-access-24vzb\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.908606 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.908625 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.908675 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.908706 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.908728 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-logs\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.908769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.908808 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.909527 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-logs\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.909761 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.910021 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.914869 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.917922 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.929051 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.930072 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.932947 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24vzb\" (UniqueName: \"kubernetes.io/projected/94883722-c05e-43f6-ac7d-91ce20de293b-kube-api-access-24vzb\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:51 crc kubenswrapper[4770]: I1004 03:21:51.935215 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.005974 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.153769 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b0c9-account-create-dx2vq" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.215551 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frm4p\" (UniqueName: \"kubernetes.io/projected/9c51889c-6112-4c86-93f1-a69876016f4f-kube-api-access-frm4p\") pod \"9c51889c-6112-4c86-93f1-a69876016f4f\" (UID: \"9c51889c-6112-4c86-93f1-a69876016f4f\") " Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.228028 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c51889c-6112-4c86-93f1-a69876016f4f-kube-api-access-frm4p" (OuterVolumeSpecName: "kube-api-access-frm4p") pod "9c51889c-6112-4c86-93f1-a69876016f4f" (UID: "9c51889c-6112-4c86-93f1-a69876016f4f"). InnerVolumeSpecName "kube-api-access-frm4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.317930 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frm4p\" (UniqueName: \"kubernetes.io/projected/9c51889c-6112-4c86-93f1-a69876016f4f-kube-api-access-frm4p\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.462486 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5365-account-create-47djg" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.526988 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc855\" (UniqueName: \"kubernetes.io/projected/483a5eae-40e5-4e11-82e8-24198ec8a733-kube-api-access-bc855\") pod \"483a5eae-40e5-4e11-82e8-24198ec8a733\" (UID: \"483a5eae-40e5-4e11-82e8-24198ec8a733\") " Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.538053 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/483a5eae-40e5-4e11-82e8-24198ec8a733-kube-api-access-bc855" (OuterVolumeSpecName: "kube-api-access-bc855") pod "483a5eae-40e5-4e11-82e8-24198ec8a733" (UID: "483a5eae-40e5-4e11-82e8-24198ec8a733"). InnerVolumeSpecName "kube-api-access-bc855". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.564218 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.578837 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5365-account-create-47djg" event={"ID":"483a5eae-40e5-4e11-82e8-24198ec8a733","Type":"ContainerDied","Data":"ccc6e88fd2c962bea367391e6ce84a3173a80e4602ffe474e9506ad7cae54e8e"} Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.578884 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccc6e88fd2c962bea367391e6ce84a3173a80e4602ffe474e9506ad7cae54e8e" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.579065 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5365-account-create-47djg" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.581374 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b0c9-account-create-dx2vq" event={"ID":"9c51889c-6112-4c86-93f1-a69876016f4f","Type":"ContainerDied","Data":"2850a2d6bebae745e198bea1b48f42b4aa3a069ad8c8562782d94fe1c5c485a2"} Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.581422 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2850a2d6bebae745e198bea1b48f42b4aa3a069ad8c8562782d94fe1c5c485a2" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.581496 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b0c9-account-create-dx2vq" Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.583661 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94883722-c05e-43f6-ac7d-91ce20de293b","Type":"ContainerStarted","Data":"469ac5cef90b33bfd29d462148a87e94a98f7d1e15639d4dec3baf0f3aacbda7"} Oct 04 03:21:52 crc kubenswrapper[4770]: I1004 03:21:52.635205 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc855\" (UniqueName: \"kubernetes.io/projected/483a5eae-40e5-4e11-82e8-24198ec8a733-kube-api-access-bc855\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:53 crc kubenswrapper[4770]: I1004 03:21:53.640320 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94883722-c05e-43f6-ac7d-91ce20de293b","Type":"ContainerStarted","Data":"36527a07ae6792b10bdaf728b57e11696b53baf86da649a634ef2e860974b76b"} Oct 04 03:21:53 crc kubenswrapper[4770]: I1004 03:21:53.671278 4770 generic.go:334] "Generic (PLEG): container finished" podID="ec2a2801-f563-4bbf-b191-7e4dffbe89a2" containerID="c2e39f05e83ac887adbf3376bdea8800dc2e57fa7efb4f00b67adc9235d689e5" exitCode=0 Oct 04 03:21:53 crc kubenswrapper[4770]: I1004 03:21:53.671356 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6886d" event={"ID":"ec2a2801-f563-4bbf-b191-7e4dffbe89a2","Type":"ContainerDied","Data":"c2e39f05e83ac887adbf3376bdea8800dc2e57fa7efb4f00b67adc9235d689e5"} Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.683074 4770 generic.go:334] "Generic (PLEG): container finished" podID="c9dfa4b6-223e-447d-b62c-1f772c82802d" containerID="df651e173b6556cb139982e461ff9c637c9cf1e2459276d87df36652f29f1557" exitCode=0 Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.683457 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nq9d8" event={"ID":"c9dfa4b6-223e-447d-b62c-1f772c82802d","Type":"ContainerDied","Data":"df651e173b6556cb139982e461ff9c637c9cf1e2459276d87df36652f29f1557"} Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.691455 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94883722-c05e-43f6-ac7d-91ce20de293b","Type":"ContainerStarted","Data":"9d13162cc33717d452f99db8ca695a367a80b9d5bc0d24acc6684388fef67953"} Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.733296 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.73326068 podStartE2EDuration="3.73326068s" podCreationTimestamp="2025-10-04 03:21:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:21:54.721848832 +0000 UTC m=+1126.013858544" watchObservedRunningTime="2025-10-04 03:21:54.73326068 +0000 UTC m=+1126.025270392" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.884043 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-6qh86"] Oct 04 03:21:54 crc kubenswrapper[4770]: E1004 03:21:54.884691 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483a5eae-40e5-4e11-82e8-24198ec8a733" containerName="mariadb-account-create" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.884711 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="483a5eae-40e5-4e11-82e8-24198ec8a733" containerName="mariadb-account-create" Oct 04 03:21:54 crc kubenswrapper[4770]: E1004 03:21:54.884735 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c51889c-6112-4c86-93f1-a69876016f4f" containerName="mariadb-account-create" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.884742 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c51889c-6112-4c86-93f1-a69876016f4f" containerName="mariadb-account-create" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.884895 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="483a5eae-40e5-4e11-82e8-24198ec8a733" containerName="mariadb-account-create" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.884921 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c51889c-6112-4c86-93f1-a69876016f4f" containerName="mariadb-account-create" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.885478 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.890469 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8l96n" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.890866 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.907268 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6qh86"] Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.997148 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-db-sync-config-data\") pod \"barbican-db-sync-6qh86\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.997198 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-combined-ca-bundle\") pod \"barbican-db-sync-6qh86\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:54 crc kubenswrapper[4770]: I1004 03:21:54.997235 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl7m2\" (UniqueName: \"kubernetes.io/projected/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-kube-api-access-gl7m2\") pod \"barbican-db-sync-6qh86\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.099512 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl7m2\" (UniqueName: \"kubernetes.io/projected/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-kube-api-access-gl7m2\") pod \"barbican-db-sync-6qh86\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.100411 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-db-sync-config-data\") pod \"barbican-db-sync-6qh86\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.100443 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-hj2s4"] Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.100488 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-combined-ca-bundle\") pod \"barbican-db-sync-6qh86\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.099671 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6886d" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.101573 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec2a2801-f563-4bbf-b191-7e4dffbe89a2" containerName="placement-db-sync" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.103062 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.109442 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.109653 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fzp9h" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.109719 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-db-sync-config-data\") pod \"barbican-db-sync-6qh86\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.109846 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.110844 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-combined-ca-bundle\") pod \"barbican-db-sync-6qh86\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.118612 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hj2s4"] Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.124416 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl7m2\" (UniqueName: \"kubernetes.io/projected/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-kube-api-access-gl7m2\") pod \"barbican-db-sync-6qh86\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.203937 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-logs\") pod \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.204274 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr6hw\" (UniqueName: \"kubernetes.io/projected/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-kube-api-access-xr6hw\") pod \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.204327 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-config-data\") pod \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.204361 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-scripts\") pod \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.204421 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-combined-ca-bundle\") pod \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\" (UID: \"ec2a2801-f563-4bbf-b191-7e4dffbe89a2\") " Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.204777 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-combined-ca-bundle\") pod \"neutron-db-sync-hj2s4\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.204822 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjn7h\" (UniqueName: \"kubernetes.io/projected/b736169c-f680-43f5-b088-110ec7e664d8-kube-api-access-pjn7h\") pod \"neutron-db-sync-hj2s4\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.204943 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-config\") pod \"neutron-db-sync-hj2s4\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.205304 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-logs" (OuterVolumeSpecName: "logs") pod "ec2a2801-f563-4bbf-b191-7e4dffbe89a2" (UID: "ec2a2801-f563-4bbf-b191-7e4dffbe89a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.209958 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-kube-api-access-xr6hw" (OuterVolumeSpecName: "kube-api-access-xr6hw") pod "ec2a2801-f563-4bbf-b191-7e4dffbe89a2" (UID: "ec2a2801-f563-4bbf-b191-7e4dffbe89a2"). InnerVolumeSpecName "kube-api-access-xr6hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.214239 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-scripts" (OuterVolumeSpecName: "scripts") pod "ec2a2801-f563-4bbf-b191-7e4dffbe89a2" (UID: "ec2a2801-f563-4bbf-b191-7e4dffbe89a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.214609 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6qh86" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.245238 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-config-data" (OuterVolumeSpecName: "config-data") pod "ec2a2801-f563-4bbf-b191-7e4dffbe89a2" (UID: "ec2a2801-f563-4bbf-b191-7e4dffbe89a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.255162 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec2a2801-f563-4bbf-b191-7e4dffbe89a2" (UID: "ec2a2801-f563-4bbf-b191-7e4dffbe89a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.306663 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-combined-ca-bundle\") pod \"neutron-db-sync-hj2s4\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.306762 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjn7h\" (UniqueName: \"kubernetes.io/projected/b736169c-f680-43f5-b088-110ec7e664d8-kube-api-access-pjn7h\") pod \"neutron-db-sync-hj2s4\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.306969 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-config\") pod \"neutron-db-sync-hj2s4\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.307113 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr6hw\" (UniqueName: \"kubernetes.io/projected/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-kube-api-access-xr6hw\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.307134 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.307152 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.307162 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.307173 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec2a2801-f563-4bbf-b191-7e4dffbe89a2-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.331621 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-config\") pod \"neutron-db-sync-hj2s4\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.359366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjn7h\" (UniqueName: \"kubernetes.io/projected/b736169c-f680-43f5-b088-110ec7e664d8-kube-api-access-pjn7h\") pod \"neutron-db-sync-hj2s4\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.411142 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-combined-ca-bundle\") pod \"neutron-db-sync-hj2s4\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.521760 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.724400 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6886d" event={"ID":"ec2a2801-f563-4bbf-b191-7e4dffbe89a2","Type":"ContainerDied","Data":"b8680169542392a169d98c36fd81c8b4071374105a56218cb2029e71339b8154"} Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.724447 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8680169542392a169d98c36fd81c8b4071374105a56218cb2029e71339b8154" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.725100 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6886d" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.827580 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-dbfd75b46-z42g2"] Oct 04 03:21:55 crc kubenswrapper[4770]: E1004 03:21:55.828259 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec2a2801-f563-4bbf-b191-7e4dffbe89a2" containerName="placement-db-sync" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.828284 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec2a2801-f563-4bbf-b191-7e4dffbe89a2" containerName="placement-db-sync" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.829621 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.834241 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9vxhs" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.835791 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.835972 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.836856 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.837271 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.863611 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-dbfd75b46-z42g2"] Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.934681 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-config-data\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.934752 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-internal-tls-certs\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.934810 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f9865e-be12-4821-a088-956061126f7b-logs\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.934834 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-public-tls-certs\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.934869 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-scripts\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.934915 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-797p7\" (UniqueName: \"kubernetes.io/projected/76f9865e-be12-4821-a088-956061126f7b-kube-api-access-797p7\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.934933 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-combined-ca-bundle\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:55 crc kubenswrapper[4770]: I1004 03:21:55.979732 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6qh86"] Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.037277 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-scripts\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.037385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-797p7\" (UniqueName: \"kubernetes.io/projected/76f9865e-be12-4821-a088-956061126f7b-kube-api-access-797p7\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.037425 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-combined-ca-bundle\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.037467 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-config-data\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.037504 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-internal-tls-certs\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.037552 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f9865e-be12-4821-a088-956061126f7b-logs\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.037581 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-public-tls-certs\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.039612 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f9865e-be12-4821-a088-956061126f7b-logs\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.049208 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-internal-tls-certs\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.049557 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-public-tls-certs\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.051481 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-scripts\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.051767 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-config-data\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.052033 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-combined-ca-bundle\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.071303 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-797p7\" (UniqueName: \"kubernetes.io/projected/76f9865e-be12-4821-a088-956061126f7b-kube-api-access-797p7\") pod \"placement-dbfd75b46-z42g2\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:56 crc kubenswrapper[4770]: I1004 03:21:56.154509 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:21:59 crc kubenswrapper[4770]: W1004 03:21:59.836688 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c0eb670_7851_4fdc_92ce_d32bcb5e79e9.slice/crio-ff8f4005bb40cc02fa1bbb78e41f34d6e9717d3914804ceee359da6bbf0178cd WatchSource:0}: Error finding container ff8f4005bb40cc02fa1bbb78e41f34d6e9717d3914804ceee359da6bbf0178cd: Status 404 returned error can't find the container with id ff8f4005bb40cc02fa1bbb78e41f34d6e9717d3914804ceee359da6bbf0178cd Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.046771 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.143041 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-fernet-keys\") pod \"c9dfa4b6-223e-447d-b62c-1f772c82802d\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.143289 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-combined-ca-bundle\") pod \"c9dfa4b6-223e-447d-b62c-1f772c82802d\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.143456 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd2nz\" (UniqueName: \"kubernetes.io/projected/c9dfa4b6-223e-447d-b62c-1f772c82802d-kube-api-access-jd2nz\") pod \"c9dfa4b6-223e-447d-b62c-1f772c82802d\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.143480 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-config-data\") pod \"c9dfa4b6-223e-447d-b62c-1f772c82802d\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.143687 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-scripts\") pod \"c9dfa4b6-223e-447d-b62c-1f772c82802d\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.143750 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-credential-keys\") pod \"c9dfa4b6-223e-447d-b62c-1f772c82802d\" (UID: \"c9dfa4b6-223e-447d-b62c-1f772c82802d\") " Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.149750 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c9dfa4b6-223e-447d-b62c-1f772c82802d" (UID: "c9dfa4b6-223e-447d-b62c-1f772c82802d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.150423 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-scripts" (OuterVolumeSpecName: "scripts") pod "c9dfa4b6-223e-447d-b62c-1f772c82802d" (UID: "c9dfa4b6-223e-447d-b62c-1f772c82802d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.151529 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9dfa4b6-223e-447d-b62c-1f772c82802d-kube-api-access-jd2nz" (OuterVolumeSpecName: "kube-api-access-jd2nz") pod "c9dfa4b6-223e-447d-b62c-1f772c82802d" (UID: "c9dfa4b6-223e-447d-b62c-1f772c82802d"). InnerVolumeSpecName "kube-api-access-jd2nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.152504 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c9dfa4b6-223e-447d-b62c-1f772c82802d" (UID: "c9dfa4b6-223e-447d-b62c-1f772c82802d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.184614 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9dfa4b6-223e-447d-b62c-1f772c82802d" (UID: "c9dfa4b6-223e-447d-b62c-1f772c82802d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.196670 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-config-data" (OuterVolumeSpecName: "config-data") pod "c9dfa4b6-223e-447d-b62c-1f772c82802d" (UID: "c9dfa4b6-223e-447d-b62c-1f772c82802d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.246060 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.246099 4770 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.246111 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.246121 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.246132 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd2nz\" (UniqueName: \"kubernetes.io/projected/c9dfa4b6-223e-447d-b62c-1f772c82802d-kube-api-access-jd2nz\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.246141 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9dfa4b6-223e-447d-b62c-1f772c82802d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.446821 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hj2s4"] Oct 04 03:22:00 crc kubenswrapper[4770]: W1004 03:22:00.449304 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb736169c_f680_43f5_b088_110ec7e664d8.slice/crio-c554c7d43bec84c27fee44be9060e0bdaae1037ab84e97e17abebcf6667326c2 WatchSource:0}: Error finding container c554c7d43bec84c27fee44be9060e0bdaae1037ab84e97e17abebcf6667326c2: Status 404 returned error can't find the container with id c554c7d43bec84c27fee44be9060e0bdaae1037ab84e97e17abebcf6667326c2 Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.622092 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-dbfd75b46-z42g2"] Oct 04 03:22:00 crc kubenswrapper[4770]: W1004 03:22:00.625434 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76f9865e_be12_4821_a088_956061126f7b.slice/crio-b6021757cffcd01c59af1172ba928c987a4ca81d21ca51c80b67cb2beed40a6e WatchSource:0}: Error finding container b6021757cffcd01c59af1172ba928c987a4ca81d21ca51c80b67cb2beed40a6e: Status 404 returned error can't find the container with id b6021757cffcd01c59af1172ba928c987a4ca81d21ca51c80b67cb2beed40a6e Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.783260 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerStarted","Data":"3f1f0d86461cea3972c663a504db410e49e384d355fe3fe09fe51879cae2010a"} Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.786610 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nq9d8" event={"ID":"c9dfa4b6-223e-447d-b62c-1f772c82802d","Type":"ContainerDied","Data":"2870179140fb8c592ec64ae32730710b3f2e5757309305b952ba6f4faea1cdad"} Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.786677 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2870179140fb8c592ec64ae32730710b3f2e5757309305b952ba6f4faea1cdad" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.786746 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nq9d8" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.790624 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dbfd75b46-z42g2" event={"ID":"76f9865e-be12-4821-a088-956061126f7b","Type":"ContainerStarted","Data":"b6021757cffcd01c59af1172ba928c987a4ca81d21ca51c80b67cb2beed40a6e"} Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.795102 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hj2s4" event={"ID":"b736169c-f680-43f5-b088-110ec7e664d8","Type":"ContainerStarted","Data":"399ec7d7f30decbc7812a69e724a27e0175509f84cc8cb66a9a7d9ac5b56cd9d"} Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.795177 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hj2s4" event={"ID":"b736169c-f680-43f5-b088-110ec7e664d8","Type":"ContainerStarted","Data":"c554c7d43bec84c27fee44be9060e0bdaae1037ab84e97e17abebcf6667326c2"} Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.798159 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6qh86" event={"ID":"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9","Type":"ContainerStarted","Data":"ff8f4005bb40cc02fa1bbb78e41f34d6e9717d3914804ceee359da6bbf0178cd"} Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.815241 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-hj2s4" podStartSLOduration=5.815211268 podStartE2EDuration="5.815211268s" podCreationTimestamp="2025-10-04 03:21:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:00.813943845 +0000 UTC m=+1132.105953557" watchObservedRunningTime="2025-10-04 03:22:00.815211268 +0000 UTC m=+1132.107220980" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.842659 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.842731 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.884429 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 03:22:00 crc kubenswrapper[4770]: I1004 03:22:00.898510 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.171123 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-866cc6bf5d-h867j"] Oct 04 03:22:01 crc kubenswrapper[4770]: E1004 03:22:01.172005 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9dfa4b6-223e-447d-b62c-1f772c82802d" containerName="keystone-bootstrap" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.172040 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9dfa4b6-223e-447d-b62c-1f772c82802d" containerName="keystone-bootstrap" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.172275 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9dfa4b6-223e-447d-b62c-1f772c82802d" containerName="keystone-bootstrap" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.172949 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.178148 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.178393 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.178666 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.178846 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wdvhv" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.178924 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.178965 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.188099 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-866cc6bf5d-h867j"] Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.268885 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-scripts\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.269060 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-fernet-keys\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.269164 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-config-data\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.269265 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-internal-tls-certs\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.269341 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcv9d\" (UniqueName: \"kubernetes.io/projected/ee078c16-5858-4c24-b936-11d4aa568f65-kube-api-access-hcv9d\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.269374 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-combined-ca-bundle\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.269607 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-credential-keys\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.269773 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-public-tls-certs\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.372677 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-public-tls-certs\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.372825 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-scripts\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.372907 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-fernet-keys\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.372969 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-config-data\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.373038 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-internal-tls-certs\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.373124 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcv9d\" (UniqueName: \"kubernetes.io/projected/ee078c16-5858-4c24-b936-11d4aa568f65-kube-api-access-hcv9d\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.373321 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-combined-ca-bundle\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.373448 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-credential-keys\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.380781 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-internal-tls-certs\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.385711 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-public-tls-certs\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.387521 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-config-data\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.387763 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-combined-ca-bundle\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.393900 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-credential-keys\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.393994 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-fernet-keys\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.396919 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-scripts\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.402258 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcv9d\" (UniqueName: \"kubernetes.io/projected/ee078c16-5858-4c24-b936-11d4aa568f65-kube-api-access-hcv9d\") pod \"keystone-866cc6bf5d-h867j\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.505213 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.813208 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dbfd75b46-z42g2" event={"ID":"76f9865e-be12-4821-a088-956061126f7b","Type":"ContainerStarted","Data":"2ed2b0c7307cdff4de6c68749e0b18dddacb71011729c291fbd02e14c6b6f756"} Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.814160 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.814203 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.814215 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dbfd75b46-z42g2" event={"ID":"76f9865e-be12-4821-a088-956061126f7b","Type":"ContainerStarted","Data":"bc0827aed61857d8364429abe1f13dfc75ffc958b4f218ff97327c98f09244f0"} Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.814244 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.814257 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 03:22:01 crc kubenswrapper[4770]: I1004 03:22:01.851397 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-dbfd75b46-z42g2" podStartSLOduration=6.851356646 podStartE2EDuration="6.851356646s" podCreationTimestamp="2025-10-04 03:21:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:01.836090646 +0000 UTC m=+1133.128100358" watchObservedRunningTime="2025-10-04 03:22:01.851356646 +0000 UTC m=+1133.143366368" Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.007217 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.007263 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.056893 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.075284 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.124829 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-866cc6bf5d-h867j"] Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.826385 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-866cc6bf5d-h867j" event={"ID":"ee078c16-5858-4c24-b936-11d4aa568f65","Type":"ContainerStarted","Data":"1f5f2c63939ced31d1a6828f3ba1395e0cbc60c93106d579d81b4050e2c43e88"} Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.828024 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.828045 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.828055 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-866cc6bf5d-h867j" event={"ID":"ee078c16-5858-4c24-b936-11d4aa568f65","Type":"ContainerStarted","Data":"96e2dfbcc9586566650da34c2453a3a3a0c2db26296fc47219d1c0dad11311c4"} Oct 04 03:22:02 crc kubenswrapper[4770]: I1004 03:22:02.857993 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-866cc6bf5d-h867j" podStartSLOduration=1.857967661 podStartE2EDuration="1.857967661s" podCreationTimestamp="2025-10-04 03:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:02.85330975 +0000 UTC m=+1134.145319462" watchObservedRunningTime="2025-10-04 03:22:02.857967661 +0000 UTC m=+1134.149977383" Oct 04 03:22:03 crc kubenswrapper[4770]: I1004 03:22:03.840183 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:04 crc kubenswrapper[4770]: I1004 03:22:04.115940 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 03:22:04 crc kubenswrapper[4770]: I1004 03:22:04.116106 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:22:04 crc kubenswrapper[4770]: I1004 03:22:04.121236 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 03:22:05 crc kubenswrapper[4770]: I1004 03:22:05.265328 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:05 crc kubenswrapper[4770]: I1004 03:22:05.265781 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:22:05 crc kubenswrapper[4770]: I1004 03:22:05.273853 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:06 crc kubenswrapper[4770]: I1004 03:22:06.908133 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6qh86" event={"ID":"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9","Type":"ContainerStarted","Data":"d7f95b4d705cde083fc4ef90c09cae5da3e5a4494a1f070421b36175c42dae52"} Oct 04 03:22:06 crc kubenswrapper[4770]: I1004 03:22:06.917720 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hkqts" event={"ID":"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a","Type":"ContainerStarted","Data":"c6a84b1e53e4e8d0fb7cc7653661bad2f904e4879e31bc7343a14c27770f22fb"} Oct 04 03:22:06 crc kubenswrapper[4770]: I1004 03:22:06.931757 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-6qh86" podStartSLOduration=6.723932662 podStartE2EDuration="12.93170313s" podCreationTimestamp="2025-10-04 03:21:54 +0000 UTC" firstStartedPulling="2025-10-04 03:21:59.883466272 +0000 UTC m=+1131.175475974" lastFinishedPulling="2025-10-04 03:22:06.09123673 +0000 UTC m=+1137.383246442" observedRunningTime="2025-10-04 03:22:06.92330744 +0000 UTC m=+1138.215317152" watchObservedRunningTime="2025-10-04 03:22:06.93170313 +0000 UTC m=+1138.223712842" Oct 04 03:22:06 crc kubenswrapper[4770]: I1004 03:22:06.955151 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-hkqts" podStartSLOduration=2.928751883 podStartE2EDuration="37.955133332s" podCreationTimestamp="2025-10-04 03:21:29 +0000 UTC" firstStartedPulling="2025-10-04 03:21:31.048134734 +0000 UTC m=+1102.340144446" lastFinishedPulling="2025-10-04 03:22:06.074516183 +0000 UTC m=+1137.366525895" observedRunningTime="2025-10-04 03:22:06.94431406 +0000 UTC m=+1138.236323792" watchObservedRunningTime="2025-10-04 03:22:06.955133332 +0000 UTC m=+1138.247143044" Oct 04 03:22:11 crc kubenswrapper[4770]: I1004 03:22:11.966466 4770 generic.go:334] "Generic (PLEG): container finished" podID="7c0eb670-7851-4fdc-92ce-d32bcb5e79e9" containerID="d7f95b4d705cde083fc4ef90c09cae5da3e5a4494a1f070421b36175c42dae52" exitCode=0 Oct 04 03:22:11 crc kubenswrapper[4770]: I1004 03:22:11.966642 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6qh86" event={"ID":"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9","Type":"ContainerDied","Data":"d7f95b4d705cde083fc4ef90c09cae5da3e5a4494a1f070421b36175c42dae52"} Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.494508 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6qh86" Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.673318 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-db-sync-config-data\") pod \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.673635 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-combined-ca-bundle\") pod \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.673820 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl7m2\" (UniqueName: \"kubernetes.io/projected/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-kube-api-access-gl7m2\") pod \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\" (UID: \"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9\") " Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.677509 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7c0eb670-7851-4fdc-92ce-d32bcb5e79e9" (UID: "7c0eb670-7851-4fdc-92ce-d32bcb5e79e9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.678091 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-kube-api-access-gl7m2" (OuterVolumeSpecName: "kube-api-access-gl7m2") pod "7c0eb670-7851-4fdc-92ce-d32bcb5e79e9" (UID: "7c0eb670-7851-4fdc-92ce-d32bcb5e79e9"). InnerVolumeSpecName "kube-api-access-gl7m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.697063 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c0eb670-7851-4fdc-92ce-d32bcb5e79e9" (UID: "7c0eb670-7851-4fdc-92ce-d32bcb5e79e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.776567 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl7m2\" (UniqueName: \"kubernetes.io/projected/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-kube-api-access-gl7m2\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.776623 4770 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.776651 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.993993 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerStarted","Data":"becc82b3357ef65b44d21bbd70ff76150feb7439d821481a05a7ba578b6fd2e5"} Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.994189 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="ceilometer-central-agent" containerID="cri-o://bb46755f85a4e4f356284e969d79b3ca1f637207f6244f543490ca472e39341a" gracePeriod=30 Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.994255 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.994255 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="sg-core" containerID="cri-o://3f1f0d86461cea3972c663a504db410e49e384d355fe3fe09fe51879cae2010a" gracePeriod=30 Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.994291 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="ceilometer-notification-agent" containerID="cri-o://88fc7604d5d5ddc8c731243d3ea5484edeab92c6728d874de648b554b94a02a8" gracePeriod=30 Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.994272 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="proxy-httpd" containerID="cri-o://becc82b3357ef65b44d21bbd70ff76150feb7439d821481a05a7ba578b6fd2e5" gracePeriod=30 Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.997721 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6qh86" event={"ID":"7c0eb670-7851-4fdc-92ce-d32bcb5e79e9","Type":"ContainerDied","Data":"ff8f4005bb40cc02fa1bbb78e41f34d6e9717d3914804ceee359da6bbf0178cd"} Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.997751 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff8f4005bb40cc02fa1bbb78e41f34d6e9717d3914804ceee359da6bbf0178cd" Oct 04 03:22:13 crc kubenswrapper[4770]: I1004 03:22:13.997795 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6qh86" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.044482 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.2034981 podStartE2EDuration="45.044457445s" podCreationTimestamp="2025-10-04 03:21:29 +0000 UTC" firstStartedPulling="2025-10-04 03:21:30.657283209 +0000 UTC m=+1101.949292921" lastFinishedPulling="2025-10-04 03:22:13.498242524 +0000 UTC m=+1144.790252266" observedRunningTime="2025-10-04 03:22:14.032761208 +0000 UTC m=+1145.324770980" watchObservedRunningTime="2025-10-04 03:22:14.044457445 +0000 UTC m=+1145.336467177" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.247078 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-66b9dbf78f-gv52v"] Oct 04 03:22:14 crc kubenswrapper[4770]: E1004 03:22:14.247577 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0eb670-7851-4fdc-92ce-d32bcb5e79e9" containerName="barbican-db-sync" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.247600 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0eb670-7851-4fdc-92ce-d32bcb5e79e9" containerName="barbican-db-sync" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.247815 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c0eb670-7851-4fdc-92ce-d32bcb5e79e9" containerName="barbican-db-sync" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.249130 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.252574 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8l96n" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.252875 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.254264 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.260386 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5877f57886-5xk9s"] Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.269205 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.270288 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5877f57886-5xk9s"] Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.278464 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-66b9dbf78f-gv52v"] Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.282077 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.388909 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data-custom\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389226 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/697bd5d5-22c6-4303-9077-434eaa099f4a-logs\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389273 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-combined-ca-bundle\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389302 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389367 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-logs\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389388 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frwv7\" (UniqueName: \"kubernetes.io/projected/697bd5d5-22c6-4303-9077-434eaa099f4a-kube-api-access-frwv7\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389402 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389426 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-combined-ca-bundle\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389440 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw4g7\" (UniqueName: \"kubernetes.io/projected/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-kube-api-access-qw4g7\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389476 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data-custom\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.389041 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd6b885b7-j2c6j"] Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.390969 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.404135 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd6b885b7-j2c6j"] Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.490730 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data-custom\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.490789 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/697bd5d5-22c6-4303-9077-434eaa099f4a-logs\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.490877 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg7ws\" (UniqueName: \"kubernetes.io/projected/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-kube-api-access-hg7ws\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.490933 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-svc\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.490990 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-combined-ca-bundle\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.491053 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.491587 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/697bd5d5-22c6-4303-9077-434eaa099f4a-logs\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493288 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-swift-storage-0\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493319 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-config\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493343 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-sb\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493387 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493507 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-logs\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493591 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frwv7\" (UniqueName: \"kubernetes.io/projected/697bd5d5-22c6-4303-9077-434eaa099f4a-kube-api-access-frwv7\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493614 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493654 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-combined-ca-bundle\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493673 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw4g7\" (UniqueName: \"kubernetes.io/projected/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-kube-api-access-qw4g7\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.493748 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data-custom\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.495917 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-logs\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.509452 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.510018 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-combined-ca-bundle\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.510350 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.510490 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-combined-ca-bundle\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.510914 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data-custom\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.520750 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frwv7\" (UniqueName: \"kubernetes.io/projected/697bd5d5-22c6-4303-9077-434eaa099f4a-kube-api-access-frwv7\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.520910 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data-custom\") pod \"barbican-worker-66b9dbf78f-gv52v\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.527462 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw4g7\" (UniqueName: \"kubernetes.io/projected/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-kube-api-access-qw4g7\") pod \"barbican-keystone-listener-5877f57886-5xk9s\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.549249 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6f69d8558d-mtg8s"] Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.551644 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.554215 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.563407 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f69d8558d-mtg8s"] Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.596375 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.599610 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg7ws\" (UniqueName: \"kubernetes.io/projected/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-kube-api-access-hg7ws\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.599660 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-svc\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.599788 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-swift-storage-0\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.599811 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-config\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.599844 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-sb\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.599879 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.600833 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.601366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-swift-storage-0\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.601533 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-config\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.601942 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-svc\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.602231 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-sb\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.621726 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg7ws\" (UniqueName: \"kubernetes.io/projected/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-kube-api-access-hg7ws\") pod \"dnsmasq-dns-7fd6b885b7-j2c6j\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.635670 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.702255 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data-custom\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.702324 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-logs\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.702365 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf6vh\" (UniqueName: \"kubernetes.io/projected/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-kube-api-access-rf6vh\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.702394 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.702417 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-combined-ca-bundle\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.783353 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.804078 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data-custom\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.804139 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-logs\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.804177 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf6vh\" (UniqueName: \"kubernetes.io/projected/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-kube-api-access-rf6vh\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.804197 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.804221 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-combined-ca-bundle\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.805480 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-logs\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.811049 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data-custom\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.811465 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.812142 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-combined-ca-bundle\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:14 crc kubenswrapper[4770]: I1004 03:22:14.853419 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf6vh\" (UniqueName: \"kubernetes.io/projected/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-kube-api-access-rf6vh\") pod \"barbican-api-6f69d8558d-mtg8s\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.106209 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.118040 4770 generic.go:334] "Generic (PLEG): container finished" podID="3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" containerID="c6a84b1e53e4e8d0fb7cc7653661bad2f904e4879e31bc7343a14c27770f22fb" exitCode=0 Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.118155 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hkqts" event={"ID":"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a","Type":"ContainerDied","Data":"c6a84b1e53e4e8d0fb7cc7653661bad2f904e4879e31bc7343a14c27770f22fb"} Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.148813 4770 generic.go:334] "Generic (PLEG): container finished" podID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerID="becc82b3357ef65b44d21bbd70ff76150feb7439d821481a05a7ba578b6fd2e5" exitCode=0 Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.148856 4770 generic.go:334] "Generic (PLEG): container finished" podID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerID="3f1f0d86461cea3972c663a504db410e49e384d355fe3fe09fe51879cae2010a" exitCode=2 Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.148864 4770 generic.go:334] "Generic (PLEG): container finished" podID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerID="88fc7604d5d5ddc8c731243d3ea5484edeab92c6728d874de648b554b94a02a8" exitCode=0 Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.148871 4770 generic.go:334] "Generic (PLEG): container finished" podID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerID="bb46755f85a4e4f356284e969d79b3ca1f637207f6244f543490ca472e39341a" exitCode=0 Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.148896 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerDied","Data":"becc82b3357ef65b44d21bbd70ff76150feb7439d821481a05a7ba578b6fd2e5"} Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.148928 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerDied","Data":"3f1f0d86461cea3972c663a504db410e49e384d355fe3fe09fe51879cae2010a"} Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.148941 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerDied","Data":"88fc7604d5d5ddc8c731243d3ea5484edeab92c6728d874de648b554b94a02a8"} Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.148951 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerDied","Data":"bb46755f85a4e4f356284e969d79b3ca1f637207f6244f543490ca472e39341a"} Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.266785 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-66b9dbf78f-gv52v"] Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.280675 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5877f57886-5xk9s"] Oct 04 03:22:15 crc kubenswrapper[4770]: W1004 03:22:15.287677 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode328a4dc_7a4d_4cb0_82a8_f60df78328d8.slice/crio-9cf55162da9e867dc8a8cde9e48d10d912ac61a3d2d08bbaa7e2df956c51e55b WatchSource:0}: Error finding container 9cf55162da9e867dc8a8cde9e48d10d912ac61a3d2d08bbaa7e2df956c51e55b: Status 404 returned error can't find the container with id 9cf55162da9e867dc8a8cde9e48d10d912ac61a3d2d08bbaa7e2df956c51e55b Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.533143 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd6b885b7-j2c6j"] Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.629806 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.649865 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-sg-core-conf-yaml\") pod \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.650039 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pj5t\" (UniqueName: \"kubernetes.io/projected/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-kube-api-access-4pj5t\") pod \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.650139 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-scripts\") pod \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.650348 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-run-httpd\") pod \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.651048 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-log-httpd\") pod \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.651100 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-config-data\") pod \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.651233 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-combined-ca-bundle\") pod \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\" (UID: \"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf\") " Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.658691 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" (UID: "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.661494 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" (UID: "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.661795 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-kube-api-access-4pj5t" (OuterVolumeSpecName: "kube-api-access-4pj5t") pod "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" (UID: "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf"). InnerVolumeSpecName "kube-api-access-4pj5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.662246 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-scripts" (OuterVolumeSpecName: "scripts") pod "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" (UID: "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.689855 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" (UID: "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:15 crc kubenswrapper[4770]: W1004 03:22:15.730469 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21e58a5a_78a0_4e87_b3ff_5d789c437dd1.slice/crio-7d918b4e978c7c4ffe7dc40f7ce0ad21b0971dddf7167f02d2d5f1fb3c05af9f WatchSource:0}: Error finding container 7d918b4e978c7c4ffe7dc40f7ce0ad21b0971dddf7167f02d2d5f1fb3c05af9f: Status 404 returned error can't find the container with id 7d918b4e978c7c4ffe7dc40f7ce0ad21b0971dddf7167f02d2d5f1fb3c05af9f Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.732377 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f69d8558d-mtg8s"] Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.757948 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.757985 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.757996 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pj5t\" (UniqueName: \"kubernetes.io/projected/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-kube-api-access-4pj5t\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.758020 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.758029 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.787417 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" (UID: "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.846142 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-config-data" (OuterVolumeSpecName: "config-data") pod "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" (UID: "9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.859540 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:15 crc kubenswrapper[4770]: I1004 03:22:15.859571 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.164474 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69d8558d-mtg8s" event={"ID":"21e58a5a-78a0-4e87-b3ff-5d789c437dd1","Type":"ContainerStarted","Data":"5db9161c1905e1df316b626cd2b33a4cb532dabb24d808c7b021a0e22baa50ef"} Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.164542 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69d8558d-mtg8s" event={"ID":"21e58a5a-78a0-4e87-b3ff-5d789c437dd1","Type":"ContainerStarted","Data":"7d918b4e978c7c4ffe7dc40f7ce0ad21b0971dddf7167f02d2d5f1fb3c05af9f"} Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.167494 4770 generic.go:334] "Generic (PLEG): container finished" podID="7c3a9c7c-9aa1-44a6-8531-7ae45343d756" containerID="46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7" exitCode=0 Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.167562 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" event={"ID":"7c3a9c7c-9aa1-44a6-8531-7ae45343d756","Type":"ContainerDied","Data":"46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7"} Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.167590 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" event={"ID":"7c3a9c7c-9aa1-44a6-8531-7ae45343d756","Type":"ContainerStarted","Data":"3aee3751e695d106f6b6ebdb8e5de8a969b00a855fa4626487bc786945af5b74"} Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.175447 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf","Type":"ContainerDied","Data":"e8de094b904b664a55471c569a961f61cd5c3bd12b3e7a876d52f3b0987800fc"} Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.175533 4770 scope.go:117] "RemoveContainer" containerID="becc82b3357ef65b44d21bbd70ff76150feb7439d821481a05a7ba578b6fd2e5" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.175476 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.182028 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66b9dbf78f-gv52v" event={"ID":"697bd5d5-22c6-4303-9077-434eaa099f4a","Type":"ContainerStarted","Data":"4c6aed319f7675ffc676d7b125c8d46484cd776ab6d78a3724fd77ab7b9f489b"} Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.184950 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" event={"ID":"e328a4dc-7a4d-4cb0-82a8-f60df78328d8","Type":"ContainerStarted","Data":"9cf55162da9e867dc8a8cde9e48d10d912ac61a3d2d08bbaa7e2df956c51e55b"} Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.216448 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.230355 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.245898 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:16 crc kubenswrapper[4770]: E1004 03:22:16.246511 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="ceilometer-notification-agent" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.246535 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="ceilometer-notification-agent" Oct 04 03:22:16 crc kubenswrapper[4770]: E1004 03:22:16.246565 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="ceilometer-central-agent" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.246574 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="ceilometer-central-agent" Oct 04 03:22:16 crc kubenswrapper[4770]: E1004 03:22:16.246601 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="proxy-httpd" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.246609 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="proxy-httpd" Oct 04 03:22:16 crc kubenswrapper[4770]: E1004 03:22:16.246624 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="sg-core" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.246631 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="sg-core" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.246845 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="sg-core" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.246867 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="ceilometer-notification-agent" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.246877 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="proxy-httpd" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.246895 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" containerName="ceilometer-central-agent" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.248959 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.253922 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.254236 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.260754 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.366535 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-log-httpd\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.366614 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-run-httpd\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.366744 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-scripts\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.367048 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-config-data\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.367104 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr55g\" (UniqueName: \"kubernetes.io/projected/3626db89-39a8-40b8-ae68-e900090e76e9-kube-api-access-nr55g\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.367243 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.367302 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.469209 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-config-data\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.469611 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr55g\" (UniqueName: \"kubernetes.io/projected/3626db89-39a8-40b8-ae68-e900090e76e9-kube-api-access-nr55g\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.469673 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.469700 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.469741 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-log-httpd\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.469764 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-run-httpd\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.469790 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-scripts\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.470795 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-run-httpd\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.470854 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-log-httpd\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.478952 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.480130 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-config-data\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.482776 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.490619 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-scripts\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.494825 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr55g\" (UniqueName: \"kubernetes.io/projected/3626db89-39a8-40b8-ae68-e900090e76e9-kube-api-access-nr55g\") pod \"ceilometer-0\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.584563 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:16 crc kubenswrapper[4770]: I1004 03:22:16.977406 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hkqts" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.079558 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-scripts\") pod \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.079602 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-config-data\") pod \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.079752 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-etc-machine-id\") pod \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.079802 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dck9k\" (UniqueName: \"kubernetes.io/projected/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-kube-api-access-dck9k\") pod \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.079832 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-db-sync-config-data\") pod \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.079900 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-combined-ca-bundle\") pod \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\" (UID: \"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a\") " Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.079894 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" (UID: "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.083840 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-scripts" (OuterVolumeSpecName: "scripts") pod "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" (UID: "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.084986 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" (UID: "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.087859 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-kube-api-access-dck9k" (OuterVolumeSpecName: "kube-api-access-dck9k") pod "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" (UID: "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a"). InnerVolumeSpecName "kube-api-access-dck9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.113042 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" (UID: "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.134885 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-config-data" (OuterVolumeSpecName: "config-data") pod "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" (UID: "3a8836e2-0ea8-43cb-ba3f-535b106c4f3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.183691 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.183735 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.183747 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.183758 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.183772 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dck9k\" (UniqueName: \"kubernetes.io/projected/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-kube-api-access-dck9k\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.183784 4770 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.199156 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hkqts" event={"ID":"3a8836e2-0ea8-43cb-ba3f-535b106c4f3a","Type":"ContainerDied","Data":"548c9e38e8dd3d3ead2d773d631f6995aed01505bf068f91b52fc28517698384"} Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.199202 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="548c9e38e8dd3d3ead2d773d631f6995aed01505bf068f91b52fc28517698384" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.199177 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hkqts" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.203215 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69d8558d-mtg8s" event={"ID":"21e58a5a-78a0-4e87-b3ff-5d789c437dd1","Type":"ContainerStarted","Data":"02071bc128c4c599cd4169231a0d430b639afeec180d3c1bb5d18d6093bf2089"} Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.203553 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.226758 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6f69d8558d-mtg8s" podStartSLOduration=3.226737979 podStartE2EDuration="3.226737979s" podCreationTimestamp="2025-10-04 03:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:17.21912265 +0000 UTC m=+1148.511132362" watchObservedRunningTime="2025-10-04 03:22:17.226737979 +0000 UTC m=+1148.518747691" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.308671 4770 scope.go:117] "RemoveContainer" containerID="3f1f0d86461cea3972c663a504db410e49e384d355fe3fe09fe51879cae2010a" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.393325 4770 scope.go:117] "RemoveContainer" containerID="88fc7604d5d5ddc8c731243d3ea5484edeab92c6728d874de648b554b94a02a8" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.592064 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:22:17 crc kubenswrapper[4770]: E1004 03:22:17.592786 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" containerName="cinder-db-sync" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.592797 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" containerName="cinder-db-sync" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.593546 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" containerName="cinder-db-sync" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.594637 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.605382 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.605591 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.605785 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.605969 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gpljl" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.606557 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.635485 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd6b885b7-j2c6j"] Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.676580 4770 scope.go:117] "RemoveContainer" containerID="bb46755f85a4e4f356284e969d79b3ca1f637207f6244f543490ca472e39341a" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.676788 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7889fffdf7-snw87"] Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.680449 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.709927 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf" path="/var/lib/kubelet/pods/9ab92fc7-3846-42d3-a6e9-ba40de3bbfbf/volumes" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.716123 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7889fffdf7-snw87"] Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.739598 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83d17d63-8752-4032-a8c1-a0d1770372e8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.739691 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.739751 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-swift-storage-0\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.739778 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-scripts\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.739817 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-nb\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.739852 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.739894 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-svc\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.739923 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj5tt\" (UniqueName: \"kubernetes.io/projected/83d17d63-8752-4032-a8c1-a0d1770372e8-kube-api-access-nj5tt\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.739999 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-sb\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.740040 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.740066 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prrwq\" (UniqueName: \"kubernetes.io/projected/ba829814-c679-46a0-99ed-d7a04e2bf96d-kube-api-access-prrwq\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.740111 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-config\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.839031 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.841778 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842590 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-swift-storage-0\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842649 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-scripts\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842716 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-nb\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842757 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842789 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-svc\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj5tt\" (UniqueName: \"kubernetes.io/projected/83d17d63-8752-4032-a8c1-a0d1770372e8-kube-api-access-nj5tt\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842863 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-sb\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842882 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842910 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prrwq\" (UniqueName: \"kubernetes.io/projected/ba829814-c679-46a0-99ed-d7a04e2bf96d-kube-api-access-prrwq\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842934 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-config\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.842975 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83d17d63-8752-4032-a8c1-a0d1770372e8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.843001 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.844506 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-swift-storage-0\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.846687 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-svc\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.847293 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-nb\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.848737 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-config\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.848803 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83d17d63-8752-4032-a8c1-a0d1770372e8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.850313 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-sb\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.850771 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.866605 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-scripts\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.866695 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.867299 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.868528 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj5tt\" (UniqueName: \"kubernetes.io/projected/83d17d63-8752-4032-a8c1-a0d1770372e8-kube-api-access-nj5tt\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.869196 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.876613 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prrwq\" (UniqueName: \"kubernetes.io/projected/ba829814-c679-46a0-99ed-d7a04e2bf96d-kube-api-access-prrwq\") pod \"dnsmasq-dns-7889fffdf7-snw87\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.889717 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.944821 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-scripts\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.944907 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-logs\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.945076 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.945108 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h6ck\" (UniqueName: \"kubernetes.io/projected/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-kube-api-access-8h6ck\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.945246 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.945386 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data-custom\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:17 crc kubenswrapper[4770]: I1004 03:22:17.945419 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.025300 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.048090 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.048245 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.048466 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h6ck\" (UniqueName: \"kubernetes.io/projected/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-kube-api-access-8h6ck\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.048719 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.048846 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data-custom\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.049080 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.049225 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-scripts\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.049374 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-logs\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.051252 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-logs\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.054293 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.054693 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-scripts\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.054970 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.055536 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data-custom\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.059452 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.073359 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.077642 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h6ck\" (UniqueName: \"kubernetes.io/projected/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-kube-api-access-8h6ck\") pod \"cinder-api-0\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.249656 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" event={"ID":"7c3a9c7c-9aa1-44a6-8531-7ae45343d756","Type":"ContainerStarted","Data":"62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309"} Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.249889 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" podUID="7c3a9c7c-9aa1-44a6-8531-7ae45343d756" containerName="dnsmasq-dns" containerID="cri-o://62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309" gracePeriod=10 Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.250456 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.301607 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.301614 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerStarted","Data":"aec713cc7199ab192a6b063d2ec83799e2c6f6d42c114a64457468efbbfd25cf"} Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.305274 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66b9dbf78f-gv52v" event={"ID":"697bd5d5-22c6-4303-9077-434eaa099f4a","Type":"ContainerStarted","Data":"ad1c6f5e9093316bec8bf19c97c384353c2014b68fb11557c892fae059e000b7"} Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.346617 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" podStartSLOduration=4.346591598 podStartE2EDuration="4.346591598s" podCreationTimestamp="2025-10-04 03:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:18.288363165 +0000 UTC m=+1149.580372877" watchObservedRunningTime="2025-10-04 03:22:18.346591598 +0000 UTC m=+1149.638601310" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.351782 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" event={"ID":"e328a4dc-7a4d-4cb0-82a8-f60df78328d8","Type":"ContainerStarted","Data":"ba22ef47e0582d9695bafcc9660a1b4adb815a65c5ef9b6fd33f435fbfb67a12"} Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.351854 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" event={"ID":"e328a4dc-7a4d-4cb0-82a8-f60df78328d8","Type":"ContainerStarted","Data":"3f304c4c9d4084df4db4c9d1ca2bc9d44285df6bb49699c82b5e8823e8e1ae3f"} Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.351885 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.376471 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" podStartSLOduration=2.293272619 podStartE2EDuration="4.376451329s" podCreationTimestamp="2025-10-04 03:22:14 +0000 UTC" firstStartedPulling="2025-10-04 03:22:15.291953302 +0000 UTC m=+1146.583963014" lastFinishedPulling="2025-10-04 03:22:17.375132012 +0000 UTC m=+1148.667141724" observedRunningTime="2025-10-04 03:22:18.369951969 +0000 UTC m=+1149.661961681" watchObservedRunningTime="2025-10-04 03:22:18.376451329 +0000 UTC m=+1149.668461051" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.381460 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-66b9dbf78f-gv52v" podStartSLOduration=2.271805786 podStartE2EDuration="4.381446469s" podCreationTimestamp="2025-10-04 03:22:14 +0000 UTC" firstStartedPulling="2025-10-04 03:22:15.286522059 +0000 UTC m=+1146.578531771" lastFinishedPulling="2025-10-04 03:22:17.396162742 +0000 UTC m=+1148.688172454" observedRunningTime="2025-10-04 03:22:18.333765802 +0000 UTC m=+1149.625775534" watchObservedRunningTime="2025-10-04 03:22:18.381446469 +0000 UTC m=+1149.673456181" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.642737 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7889fffdf7-snw87"] Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.776455 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:22:18 crc kubenswrapper[4770]: W1004 03:22:18.819414 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83d17d63_8752_4032_a8c1_a0d1770372e8.slice/crio-23acfbe2d593914a91300046a9cca755f0b62c2b21688010078e392feb88e500 WatchSource:0}: Error finding container 23acfbe2d593914a91300046a9cca755f0b62c2b21688010078e392feb88e500: Status 404 returned error can't find the container with id 23acfbe2d593914a91300046a9cca755f0b62c2b21688010078e392feb88e500 Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.848736 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.912802 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-swift-storage-0\") pod \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.913020 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-svc\") pod \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.915937 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg7ws\" (UniqueName: \"kubernetes.io/projected/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-kube-api-access-hg7ws\") pod \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.916178 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-config\") pod \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.916301 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-sb\") pod \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.916493 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-nb\") pod \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\" (UID: \"7c3a9c7c-9aa1-44a6-8531-7ae45343d756\") " Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.934176 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-kube-api-access-hg7ws" (OuterVolumeSpecName: "kube-api-access-hg7ws") pod "7c3a9c7c-9aa1-44a6-8531-7ae45343d756" (UID: "7c3a9c7c-9aa1-44a6-8531-7ae45343d756"). InnerVolumeSpecName "kube-api-access-hg7ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:18 crc kubenswrapper[4770]: I1004 03:22:18.982871 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.020442 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg7ws\" (UniqueName: \"kubernetes.io/projected/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-kube-api-access-hg7ws\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.036002 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7c3a9c7c-9aa1-44a6-8531-7ae45343d756" (UID: "7c3a9c7c-9aa1-44a6-8531-7ae45343d756"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.089671 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7c3a9c7c-9aa1-44a6-8531-7ae45343d756" (UID: "7c3a9c7c-9aa1-44a6-8531-7ae45343d756"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.089860 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7c3a9c7c-9aa1-44a6-8531-7ae45343d756" (UID: "7c3a9c7c-9aa1-44a6-8531-7ae45343d756"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.090974 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7c3a9c7c-9aa1-44a6-8531-7ae45343d756" (UID: "7c3a9c7c-9aa1-44a6-8531-7ae45343d756"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.096953 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-config" (OuterVolumeSpecName: "config") pod "7c3a9c7c-9aa1-44a6-8531-7ae45343d756" (UID: "7c3a9c7c-9aa1-44a6-8531-7ae45343d756"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.123076 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.123117 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.123128 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.123138 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.123150 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3a9c7c-9aa1-44a6-8531-7ae45343d756-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.374184 4770 generic.go:334] "Generic (PLEG): container finished" podID="7c3a9c7c-9aa1-44a6-8531-7ae45343d756" containerID="62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309" exitCode=0 Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.374287 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" event={"ID":"7c3a9c7c-9aa1-44a6-8531-7ae45343d756","Type":"ContainerDied","Data":"62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309"} Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.374338 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" event={"ID":"7c3a9c7c-9aa1-44a6-8531-7ae45343d756","Type":"ContainerDied","Data":"3aee3751e695d106f6b6ebdb8e5de8a969b00a855fa4626487bc786945af5b74"} Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.374370 4770 scope.go:117] "RemoveContainer" containerID="62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.374576 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd6b885b7-j2c6j" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.378130 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerStarted","Data":"c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c"} Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.380346 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66b9dbf78f-gv52v" event={"ID":"697bd5d5-22c6-4303-9077-434eaa099f4a","Type":"ContainerStarted","Data":"ca61c9dff3e5c232108ba39a8b1a8f4f4f5b0eb23bc5838c1ee8c7e06106319e"} Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.385782 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83d17d63-8752-4032-a8c1-a0d1770372e8","Type":"ContainerStarted","Data":"23acfbe2d593914a91300046a9cca755f0b62c2b21688010078e392feb88e500"} Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.388642 4770 generic.go:334] "Generic (PLEG): container finished" podID="ba829814-c679-46a0-99ed-d7a04e2bf96d" containerID="80067f66c1954d7df70f8381839b5e771c4dde9379cad0ef31951faee78c84e0" exitCode=0 Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.388766 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" event={"ID":"ba829814-c679-46a0-99ed-d7a04e2bf96d","Type":"ContainerDied","Data":"80067f66c1954d7df70f8381839b5e771c4dde9379cad0ef31951faee78c84e0"} Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.388787 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" event={"ID":"ba829814-c679-46a0-99ed-d7a04e2bf96d","Type":"ContainerStarted","Data":"b170c0fe1a38cb16c9349d907191b2b9febc4da2bcdd964a791e280c71fe7de0"} Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.415787 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10f8dce3-25e3-4b46-9c28-a8be0d7825ee","Type":"ContainerStarted","Data":"dbfd358a96361f106198b97df6fa76aae107db1980b1855f4c2572b0e57a1ec9"} Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.620470 4770 scope.go:117] "RemoveContainer" containerID="46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.655497 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd6b885b7-j2c6j"] Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.670609 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd6b885b7-j2c6j"] Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.694714 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3a9c7c-9aa1-44a6-8531-7ae45343d756" path="/var/lib/kubelet/pods/7c3a9c7c-9aa1-44a6-8531-7ae45343d756/volumes" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.755382 4770 scope.go:117] "RemoveContainer" containerID="62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309" Oct 04 03:22:19 crc kubenswrapper[4770]: E1004 03:22:19.758526 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309\": container with ID starting with 62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309 not found: ID does not exist" containerID="62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.758590 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309"} err="failed to get container status \"62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309\": rpc error: code = NotFound desc = could not find container \"62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309\": container with ID starting with 62eacdd47ce4379d0bc9836962c08d9e91c3783c9926c2be7394594458eef309 not found: ID does not exist" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.758628 4770 scope.go:117] "RemoveContainer" containerID="46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7" Oct 04 03:22:19 crc kubenswrapper[4770]: E1004 03:22:19.759101 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7\": container with ID starting with 46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7 not found: ID does not exist" containerID="46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7" Oct 04 03:22:19 crc kubenswrapper[4770]: I1004 03:22:19.759133 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7"} err="failed to get container status \"46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7\": rpc error: code = NotFound desc = could not find container \"46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7\": container with ID starting with 46a392924cb7e6e068615e15896c018f8fe402debd1dca2a26b433708f218bf7 not found: ID does not exist" Oct 04 03:22:20 crc kubenswrapper[4770]: I1004 03:22:20.441993 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" event={"ID":"ba829814-c679-46a0-99ed-d7a04e2bf96d","Type":"ContainerStarted","Data":"cdee7b6e2ac49a189d7a69bac5314a7d4296f6bd3b020ac64a10b6275ad7a4f6"} Oct 04 03:22:20 crc kubenswrapper[4770]: I1004 03:22:20.442614 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:20 crc kubenswrapper[4770]: I1004 03:22:20.464494 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10f8dce3-25e3-4b46-9c28-a8be0d7825ee","Type":"ContainerStarted","Data":"9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508"} Oct 04 03:22:20 crc kubenswrapper[4770]: I1004 03:22:20.470107 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" podStartSLOduration=3.470089553 podStartE2EDuration="3.470089553s" podCreationTimestamp="2025-10-04 03:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:20.463893481 +0000 UTC m=+1151.755903193" watchObservedRunningTime="2025-10-04 03:22:20.470089553 +0000 UTC m=+1151.762099265" Oct 04 03:22:20 crc kubenswrapper[4770]: I1004 03:22:20.471259 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerStarted","Data":"7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753"} Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.310679 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.494753 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerStarted","Data":"a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab"} Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.500978 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83d17d63-8752-4032-a8c1-a0d1770372e8","Type":"ContainerStarted","Data":"c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d"} Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.518457 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10f8dce3-25e3-4b46-9c28-a8be0d7825ee","Type":"ContainerStarted","Data":"39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89"} Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.519056 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.556170 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.556152488 podStartE2EDuration="4.556152488s" podCreationTimestamp="2025-10-04 03:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:21.555358376 +0000 UTC m=+1152.847368088" watchObservedRunningTime="2025-10-04 03:22:21.556152488 +0000 UTC m=+1152.848162200" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.829438 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6f85c64c7b-jcqvq"] Oct 04 03:22:21 crc kubenswrapper[4770]: E1004 03:22:21.829805 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3a9c7c-9aa1-44a6-8531-7ae45343d756" containerName="dnsmasq-dns" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.829820 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3a9c7c-9aa1-44a6-8531-7ae45343d756" containerName="dnsmasq-dns" Oct 04 03:22:21 crc kubenswrapper[4770]: E1004 03:22:21.829849 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3a9c7c-9aa1-44a6-8531-7ae45343d756" containerName="init" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.829855 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3a9c7c-9aa1-44a6-8531-7ae45343d756" containerName="init" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.830042 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3a9c7c-9aa1-44a6-8531-7ae45343d756" containerName="dnsmasq-dns" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.830879 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.838245 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.838304 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.875499 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f85c64c7b-jcqvq"] Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.898199 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-internal-tls-certs\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.898368 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-combined-ca-bundle\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.898440 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data-custom\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.898502 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.899463 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d745cd-57f5-4624-9d2c-75353e52fa56-logs\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.899587 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8v9f\" (UniqueName: \"kubernetes.io/projected/83d745cd-57f5-4624-9d2c-75353e52fa56-kube-api-access-t8v9f\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:21 crc kubenswrapper[4770]: I1004 03:22:21.899676 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-public-tls-certs\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.000945 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-combined-ca-bundle\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.001379 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data-custom\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.001442 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.001479 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d745cd-57f5-4624-9d2c-75353e52fa56-logs\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.001528 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8v9f\" (UniqueName: \"kubernetes.io/projected/83d745cd-57f5-4624-9d2c-75353e52fa56-kube-api-access-t8v9f\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.001592 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-public-tls-certs\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.001625 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-internal-tls-certs\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.002756 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d745cd-57f5-4624-9d2c-75353e52fa56-logs\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.008147 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-combined-ca-bundle\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.009726 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data-custom\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.018482 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.024927 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-internal-tls-certs\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.025210 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-public-tls-certs\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.028182 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8v9f\" (UniqueName: \"kubernetes.io/projected/83d745cd-57f5-4624-9d2c-75353e52fa56-kube-api-access-t8v9f\") pod \"barbican-api-6f85c64c7b-jcqvq\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.153704 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.538644 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerName="cinder-api-log" containerID="cri-o://9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508" gracePeriod=30 Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.539204 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83d17d63-8752-4032-a8c1-a0d1770372e8","Type":"ContainerStarted","Data":"6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075"} Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.539247 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerName="cinder-api" containerID="cri-o://39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89" gracePeriod=30 Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.798742 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.361781863 podStartE2EDuration="5.798683405s" podCreationTimestamp="2025-10-04 03:22:17 +0000 UTC" firstStartedPulling="2025-10-04 03:22:18.834624146 +0000 UTC m=+1150.126633858" lastFinishedPulling="2025-10-04 03:22:20.271525688 +0000 UTC m=+1151.563535400" observedRunningTime="2025-10-04 03:22:22.570211617 +0000 UTC m=+1153.862221329" watchObservedRunningTime="2025-10-04 03:22:22.798683405 +0000 UTC m=+1154.090693147" Oct 04 03:22:22 crc kubenswrapper[4770]: I1004 03:22:22.801272 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f85c64c7b-jcqvq"] Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.055608 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.166671 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.242404 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h6ck\" (UniqueName: \"kubernetes.io/projected/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-kube-api-access-8h6ck\") pod \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.242520 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-logs\") pod \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.242545 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-combined-ca-bundle\") pod \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.242587 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data-custom\") pod \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.242670 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data\") pod \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.242693 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-scripts\") pod \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.242716 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-etc-machine-id\") pod \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\" (UID: \"10f8dce3-25e3-4b46-9c28-a8be0d7825ee\") " Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.243165 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "10f8dce3-25e3-4b46-9c28-a8be0d7825ee" (UID: "10f8dce3-25e3-4b46-9c28-a8be0d7825ee"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.243773 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-logs" (OuterVolumeSpecName: "logs") pod "10f8dce3-25e3-4b46-9c28-a8be0d7825ee" (UID: "10f8dce3-25e3-4b46-9c28-a8be0d7825ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.253388 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "10f8dce3-25e3-4b46-9c28-a8be0d7825ee" (UID: "10f8dce3-25e3-4b46-9c28-a8be0d7825ee"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.253458 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-kube-api-access-8h6ck" (OuterVolumeSpecName: "kube-api-access-8h6ck") pod "10f8dce3-25e3-4b46-9c28-a8be0d7825ee" (UID: "10f8dce3-25e3-4b46-9c28-a8be0d7825ee"). InnerVolumeSpecName "kube-api-access-8h6ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.260288 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-scripts" (OuterVolumeSpecName: "scripts") pod "10f8dce3-25e3-4b46-9c28-a8be0d7825ee" (UID: "10f8dce3-25e3-4b46-9c28-a8be0d7825ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.284336 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10f8dce3-25e3-4b46-9c28-a8be0d7825ee" (UID: "10f8dce3-25e3-4b46-9c28-a8be0d7825ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.345516 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h6ck\" (UniqueName: \"kubernetes.io/projected/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-kube-api-access-8h6ck\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.345553 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.345566 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.345578 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.345587 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.345595 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.386105 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data" (OuterVolumeSpecName: "config-data") pod "10f8dce3-25e3-4b46-9c28-a8be0d7825ee" (UID: "10f8dce3-25e3-4b46-9c28-a8be0d7825ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.447906 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f8dce3-25e3-4b46-9c28-a8be0d7825ee-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.551934 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f85c64c7b-jcqvq" event={"ID":"83d745cd-57f5-4624-9d2c-75353e52fa56","Type":"ContainerStarted","Data":"76df273164d25e12be0bd5fba204411c763fba6be089bd2cf529d36352a6edac"} Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.551979 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f85c64c7b-jcqvq" event={"ID":"83d745cd-57f5-4624-9d2c-75353e52fa56","Type":"ContainerStarted","Data":"3fc97d22a0ca2a02e0e8c90826afa1c5686946f611beac9c19eacf4fb2e9a19d"} Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.554239 4770 generic.go:334] "Generic (PLEG): container finished" podID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerID="39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89" exitCode=0 Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.554267 4770 generic.go:334] "Generic (PLEG): container finished" podID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerID="9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508" exitCode=143 Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.554497 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10f8dce3-25e3-4b46-9c28-a8be0d7825ee","Type":"ContainerDied","Data":"39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89"} Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.554530 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10f8dce3-25e3-4b46-9c28-a8be0d7825ee","Type":"ContainerDied","Data":"9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508"} Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.554559 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10f8dce3-25e3-4b46-9c28-a8be0d7825ee","Type":"ContainerDied","Data":"dbfd358a96361f106198b97df6fa76aae107db1980b1855f4c2572b0e57a1ec9"} Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.555092 4770 scope.go:117] "RemoveContainer" containerID="39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.555266 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.583814 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerStarted","Data":"5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd"} Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.612110 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.634366 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.653788 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:22:23 crc kubenswrapper[4770]: E1004 03:22:23.655122 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerName="cinder-api-log" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.655152 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerName="cinder-api-log" Oct 04 03:22:23 crc kubenswrapper[4770]: E1004 03:22:23.655177 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerName="cinder-api" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.655186 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerName="cinder-api" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.655491 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerName="cinder-api-log" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.655514 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" containerName="cinder-api" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.655543 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.264285256 podStartE2EDuration="7.655509191s" podCreationTimestamp="2025-10-04 03:22:16 +0000 UTC" firstStartedPulling="2025-10-04 03:22:18.037350697 +0000 UTC m=+1149.329360409" lastFinishedPulling="2025-10-04 03:22:22.428574632 +0000 UTC m=+1153.720584344" observedRunningTime="2025-10-04 03:22:23.630554568 +0000 UTC m=+1154.922564280" watchObservedRunningTime="2025-10-04 03:22:23.655509191 +0000 UTC m=+1154.947518903" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.657218 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.661330 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.672939 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.673096 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.691883 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10f8dce3-25e3-4b46-9c28-a8be0d7825ee" path="/var/lib/kubelet/pods/10f8dce3-25e3-4b46-9c28-a8be0d7825ee/volumes" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.713217 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.757925 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh9vh\" (UniqueName: \"kubernetes.io/projected/87293d1f-5e33-4a9a-9978-67487fd14809-kube-api-access-gh9vh\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.758066 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87293d1f-5e33-4a9a-9978-67487fd14809-etc-machine-id\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.758157 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-public-tls-certs\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.758997 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87293d1f-5e33-4a9a-9978-67487fd14809-logs\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.759229 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-scripts\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.759283 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data-custom\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.759301 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.759419 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.759440 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.860833 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-scripts\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.860887 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data-custom\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.860904 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.860940 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.860957 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.860984 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh9vh\" (UniqueName: \"kubernetes.io/projected/87293d1f-5e33-4a9a-9978-67487fd14809-kube-api-access-gh9vh\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.861054 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87293d1f-5e33-4a9a-9978-67487fd14809-etc-machine-id\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.861114 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-public-tls-certs\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.861137 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87293d1f-5e33-4a9a-9978-67487fd14809-logs\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.861348 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87293d1f-5e33-4a9a-9978-67487fd14809-etc-machine-id\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.861718 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87293d1f-5e33-4a9a-9978-67487fd14809-logs\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.865478 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-scripts\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.866712 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.867940 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data-custom\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.873471 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-public-tls-certs\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.875599 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.877999 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:23 crc kubenswrapper[4770]: I1004 03:22:23.878792 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh9vh\" (UniqueName: \"kubernetes.io/projected/87293d1f-5e33-4a9a-9978-67487fd14809-kube-api-access-gh9vh\") pod \"cinder-api-0\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " pod="openstack/cinder-api-0" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.011486 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.218379 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6f69d8558d-mtg8s" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.605497 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.669619 4770 scope.go:117] "RemoveContainer" containerID="9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.751957 4770 scope.go:117] "RemoveContainer" containerID="39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89" Oct 04 03:22:24 crc kubenswrapper[4770]: E1004 03:22:24.752491 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89\": container with ID starting with 39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89 not found: ID does not exist" containerID="39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.752555 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89"} err="failed to get container status \"39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89\": rpc error: code = NotFound desc = could not find container \"39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89\": container with ID starting with 39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89 not found: ID does not exist" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.752589 4770 scope.go:117] "RemoveContainer" containerID="9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508" Oct 04 03:22:24 crc kubenswrapper[4770]: E1004 03:22:24.753064 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508\": container with ID starting with 9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508 not found: ID does not exist" containerID="9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.753131 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508"} err="failed to get container status \"9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508\": rpc error: code = NotFound desc = could not find container \"9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508\": container with ID starting with 9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508 not found: ID does not exist" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.753169 4770 scope.go:117] "RemoveContainer" containerID="39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.753477 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89"} err="failed to get container status \"39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89\": rpc error: code = NotFound desc = could not find container \"39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89\": container with ID starting with 39cbc7c904a927802317d0bcbcadc06e5aa7379f4c9bb826cbd7b3bc15c84c89 not found: ID does not exist" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.753517 4770 scope.go:117] "RemoveContainer" containerID="9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508" Oct 04 03:22:24 crc kubenswrapper[4770]: I1004 03:22:24.753887 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508"} err="failed to get container status \"9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508\": rpc error: code = NotFound desc = could not find container \"9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508\": container with ID starting with 9f0d2dedca1ccae0f703866e187036d03f63a93b078a7cd6d05a8c276f2ae508 not found: ID does not exist" Oct 04 03:22:25 crc kubenswrapper[4770]: I1004 03:22:25.189516 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:22:25 crc kubenswrapper[4770]: I1004 03:22:25.613950 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87293d1f-5e33-4a9a-9978-67487fd14809","Type":"ContainerStarted","Data":"a355b29233b80cd9d87bf672cfc5ce2d7d209b5e3ebe97a8765ab09f7cbb3c9e"} Oct 04 03:22:25 crc kubenswrapper[4770]: I1004 03:22:25.620739 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f85c64c7b-jcqvq" event={"ID":"83d745cd-57f5-4624-9d2c-75353e52fa56","Type":"ContainerStarted","Data":"f652f94cc4c1e2df6ea6011817891a4c91174bbbbee764ff5579bda21b7ab8fb"} Oct 04 03:22:25 crc kubenswrapper[4770]: I1004 03:22:25.620902 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:25 crc kubenswrapper[4770]: I1004 03:22:25.621027 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:25 crc kubenswrapper[4770]: I1004 03:22:25.643536 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6f85c64c7b-jcqvq" podStartSLOduration=4.643520312 podStartE2EDuration="4.643520312s" podCreationTimestamp="2025-10-04 03:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:25.641958161 +0000 UTC m=+1156.933967883" watchObservedRunningTime="2025-10-04 03:22:25.643520312 +0000 UTC m=+1156.935530024" Oct 04 03:22:26 crc kubenswrapper[4770]: I1004 03:22:26.610824 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:26 crc kubenswrapper[4770]: I1004 03:22:26.625618 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:26 crc kubenswrapper[4770]: I1004 03:22:26.632890 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87293d1f-5e33-4a9a-9978-67487fd14809","Type":"ContainerStarted","Data":"2ba074cd36ae8e2080bb5c06e842e5c1124a1cecb16c5a3f5f1ebd17d268f918"} Oct 04 03:22:26 crc kubenswrapper[4770]: I1004 03:22:26.632931 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87293d1f-5e33-4a9a-9978-67487fd14809","Type":"ContainerStarted","Data":"68b1223ff961d5213ea6bc5690e64d21567c84142a78aaf580e2d2077be0f0d2"} Oct 04 03:22:26 crc kubenswrapper[4770]: I1004 03:22:26.696699 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.696671915 podStartE2EDuration="3.696671915s" podCreationTimestamp="2025-10-04 03:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:26.688542982 +0000 UTC m=+1157.980552694" watchObservedRunningTime="2025-10-04 03:22:26.696671915 +0000 UTC m=+1157.988681627" Oct 04 03:22:27 crc kubenswrapper[4770]: I1004 03:22:27.275834 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:22:27 crc kubenswrapper[4770]: I1004 03:22:27.282443 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:22:27 crc kubenswrapper[4770]: I1004 03:22:27.643970 4770 generic.go:334] "Generic (PLEG): container finished" podID="b736169c-f680-43f5-b088-110ec7e664d8" containerID="399ec7d7f30decbc7812a69e724a27e0175509f84cc8cb66a9a7d9ac5b56cd9d" exitCode=0 Oct 04 03:22:27 crc kubenswrapper[4770]: I1004 03:22:27.644045 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hj2s4" event={"ID":"b736169c-f680-43f5-b088-110ec7e664d8","Type":"ContainerDied","Data":"399ec7d7f30decbc7812a69e724a27e0175509f84cc8cb66a9a7d9ac5b56cd9d"} Oct 04 03:22:27 crc kubenswrapper[4770]: I1004 03:22:27.644841 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.075764 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.197749 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-h4h4h"] Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.198087 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" podUID="13c7863f-832e-4a86-8ce6-f57c92e834c6" containerName="dnsmasq-dns" containerID="cri-o://0259add0c4189d05b42efd063fd05d0d7b7cec99654b19c19844b1ff27eb64c4" gracePeriod=10 Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.390360 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.439677 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:22:28 crc kubenswrapper[4770]: E1004 03:22:28.445668 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13c7863f_832e_4a86_8ce6_f57c92e834c6.slice/crio-0259add0c4189d05b42efd063fd05d0d7b7cec99654b19c19844b1ff27eb64c4.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.659565 4770 generic.go:334] "Generic (PLEG): container finished" podID="13c7863f-832e-4a86-8ce6-f57c92e834c6" containerID="0259add0c4189d05b42efd063fd05d0d7b7cec99654b19c19844b1ff27eb64c4" exitCode=0 Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.659790 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" event={"ID":"13c7863f-832e-4a86-8ce6-f57c92e834c6","Type":"ContainerDied","Data":"0259add0c4189d05b42efd063fd05d0d7b7cec99654b19c19844b1ff27eb64c4"} Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.659820 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" event={"ID":"13c7863f-832e-4a86-8ce6-f57c92e834c6","Type":"ContainerDied","Data":"151b88d45affba4f10727851e4b55864ae77a058e266690913507173a6bf5625"} Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.659841 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="151b88d45affba4f10727851e4b55864ae77a058e266690913507173a6bf5625" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.661042 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerName="cinder-scheduler" containerID="cri-o://c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d" gracePeriod=30 Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.661691 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerName="probe" containerID="cri-o://6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075" gracePeriod=30 Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.711793 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.860081 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-config\") pod \"13c7863f-832e-4a86-8ce6-f57c92e834c6\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.860470 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-svc\") pod \"13c7863f-832e-4a86-8ce6-f57c92e834c6\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.860961 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tqql\" (UniqueName: \"kubernetes.io/projected/13c7863f-832e-4a86-8ce6-f57c92e834c6-kube-api-access-7tqql\") pod \"13c7863f-832e-4a86-8ce6-f57c92e834c6\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.861036 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-swift-storage-0\") pod \"13c7863f-832e-4a86-8ce6-f57c92e834c6\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.861069 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-nb\") pod \"13c7863f-832e-4a86-8ce6-f57c92e834c6\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.861131 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-sb\") pod \"13c7863f-832e-4a86-8ce6-f57c92e834c6\" (UID: \"13c7863f-832e-4a86-8ce6-f57c92e834c6\") " Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.912112 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13c7863f-832e-4a86-8ce6-f57c92e834c6-kube-api-access-7tqql" (OuterVolumeSpecName: "kube-api-access-7tqql") pod "13c7863f-832e-4a86-8ce6-f57c92e834c6" (UID: "13c7863f-832e-4a86-8ce6-f57c92e834c6"). InnerVolumeSpecName "kube-api-access-7tqql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.942404 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "13c7863f-832e-4a86-8ce6-f57c92e834c6" (UID: "13c7863f-832e-4a86-8ce6-f57c92e834c6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.951125 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "13c7863f-832e-4a86-8ce6-f57c92e834c6" (UID: "13c7863f-832e-4a86-8ce6-f57c92e834c6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.952413 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-config" (OuterVolumeSpecName: "config") pod "13c7863f-832e-4a86-8ce6-f57c92e834c6" (UID: "13c7863f-832e-4a86-8ce6-f57c92e834c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.963182 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.963216 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tqql\" (UniqueName: \"kubernetes.io/projected/13c7863f-832e-4a86-8ce6-f57c92e834c6-kube-api-access-7tqql\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.963230 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:28 crc kubenswrapper[4770]: I1004 03:22:28.963240 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.038871 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.039452 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "13c7863f-832e-4a86-8ce6-f57c92e834c6" (UID: "13c7863f-832e-4a86-8ce6-f57c92e834c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.052782 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "13c7863f-832e-4a86-8ce6-f57c92e834c6" (UID: "13c7863f-832e-4a86-8ce6-f57c92e834c6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.064866 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.064901 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13c7863f-832e-4a86-8ce6-f57c92e834c6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.085403 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.166499 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-combined-ca-bundle\") pod \"b736169c-f680-43f5-b088-110ec7e664d8\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.166793 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-config\") pod \"b736169c-f680-43f5-b088-110ec7e664d8\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.167321 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjn7h\" (UniqueName: \"kubernetes.io/projected/b736169c-f680-43f5-b088-110ec7e664d8-kube-api-access-pjn7h\") pod \"b736169c-f680-43f5-b088-110ec7e664d8\" (UID: \"b736169c-f680-43f5-b088-110ec7e664d8\") " Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.171150 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b736169c-f680-43f5-b088-110ec7e664d8-kube-api-access-pjn7h" (OuterVolumeSpecName: "kube-api-access-pjn7h") pod "b736169c-f680-43f5-b088-110ec7e664d8" (UID: "b736169c-f680-43f5-b088-110ec7e664d8"). InnerVolumeSpecName "kube-api-access-pjn7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.171758 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjn7h\" (UniqueName: \"kubernetes.io/projected/b736169c-f680-43f5-b088-110ec7e664d8-kube-api-access-pjn7h\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.190590 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-config" (OuterVolumeSpecName: "config") pod "b736169c-f680-43f5-b088-110ec7e664d8" (UID: "b736169c-f680-43f5-b088-110ec7e664d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.190704 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b736169c-f680-43f5-b088-110ec7e664d8" (UID: "b736169c-f680-43f5-b088-110ec7e664d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.285070 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.285107 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b736169c-f680-43f5-b088-110ec7e664d8-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.671946 4770 generic.go:334] "Generic (PLEG): container finished" podID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerID="6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075" exitCode=0 Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.671999 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83d17d63-8752-4032-a8c1-a0d1770372e8","Type":"ContainerDied","Data":"6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075"} Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.687584 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hj2s4" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.687735 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6795bd975-h4h4h" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.689644 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hj2s4" event={"ID":"b736169c-f680-43f5-b088-110ec7e664d8","Type":"ContainerDied","Data":"c554c7d43bec84c27fee44be9060e0bdaae1037ab84e97e17abebcf6667326c2"} Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.689697 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c554c7d43bec84c27fee44be9060e0bdaae1037ab84e97e17abebcf6667326c2" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.743469 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-h4h4h"] Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.752403 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6795bd975-h4h4h"] Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.928912 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-zfjdv"] Oct 04 03:22:29 crc kubenswrapper[4770]: E1004 03:22:29.929967 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c7863f-832e-4a86-8ce6-f57c92e834c6" containerName="init" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.929985 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c7863f-832e-4a86-8ce6-f57c92e834c6" containerName="init" Oct 04 03:22:29 crc kubenswrapper[4770]: E1004 03:22:29.930033 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b736169c-f680-43f5-b088-110ec7e664d8" containerName="neutron-db-sync" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.930039 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b736169c-f680-43f5-b088-110ec7e664d8" containerName="neutron-db-sync" Oct 04 03:22:29 crc kubenswrapper[4770]: E1004 03:22:29.930050 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c7863f-832e-4a86-8ce6-f57c92e834c6" containerName="dnsmasq-dns" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.930057 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c7863f-832e-4a86-8ce6-f57c92e834c6" containerName="dnsmasq-dns" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.935174 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c7863f-832e-4a86-8ce6-f57c92e834c6" containerName="dnsmasq-dns" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.935217 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b736169c-f680-43f5-b088-110ec7e664d8" containerName="neutron-db-sync" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.942339 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:29 crc kubenswrapper[4770]: I1004 03:22:29.965686 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-zfjdv"] Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.051932 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5dbdffbfb6-pnbvk"] Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.055953 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.059957 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.060290 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.060961 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fzp9h" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.064459 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.108884 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-nb\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.109558 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwsgh\" (UniqueName: \"kubernetes.io/projected/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-kube-api-access-qwsgh\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.109781 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-svc\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.109953 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-config\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.110080 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-swift-storage-0\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.110388 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-sb\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.113672 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5dbdffbfb6-pnbvk"] Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.212124 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-config\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.212198 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-swift-storage-0\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213204 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-config\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213272 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rzdb\" (UniqueName: \"kubernetes.io/projected/43a28d77-6122-4774-b40d-f92dfdc9933d-kube-api-access-9rzdb\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213316 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-config\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213355 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-swift-storage-0\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213374 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-sb\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213395 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-httpd-config\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213427 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-nb\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213460 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-ovndb-tls-certs\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213547 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-combined-ca-bundle\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213588 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwsgh\" (UniqueName: \"kubernetes.io/projected/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-kube-api-access-qwsgh\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.213616 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-svc\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.214063 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-sb\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.216424 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-nb\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.216812 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-svc\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.250989 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwsgh\" (UniqueName: \"kubernetes.io/projected/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-kube-api-access-qwsgh\") pod \"dnsmasq-dns-548c5895b5-zfjdv\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.273825 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.315691 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rzdb\" (UniqueName: \"kubernetes.io/projected/43a28d77-6122-4774-b40d-f92dfdc9933d-kube-api-access-9rzdb\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.316299 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-config\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.316370 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-httpd-config\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.316419 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-ovndb-tls-certs\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.316743 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-combined-ca-bundle\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.324442 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-ovndb-tls-certs\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.326856 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-combined-ca-bundle\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.327217 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-config\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.334168 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rzdb\" (UniqueName: \"kubernetes.io/projected/43a28d77-6122-4774-b40d-f92dfdc9933d-kube-api-access-9rzdb\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.351159 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-httpd-config\") pod \"neutron-5dbdffbfb6-pnbvk\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.384403 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.417362 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.520965 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data-custom\") pod \"83d17d63-8752-4032-a8c1-a0d1770372e8\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.521070 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-combined-ca-bundle\") pod \"83d17d63-8752-4032-a8c1-a0d1770372e8\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.521153 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj5tt\" (UniqueName: \"kubernetes.io/projected/83d17d63-8752-4032-a8c1-a0d1770372e8-kube-api-access-nj5tt\") pod \"83d17d63-8752-4032-a8c1-a0d1770372e8\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.521871 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data\") pod \"83d17d63-8752-4032-a8c1-a0d1770372e8\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.521968 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83d17d63-8752-4032-a8c1-a0d1770372e8-etc-machine-id\") pod \"83d17d63-8752-4032-a8c1-a0d1770372e8\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.522027 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-scripts\") pod \"83d17d63-8752-4032-a8c1-a0d1770372e8\" (UID: \"83d17d63-8752-4032-a8c1-a0d1770372e8\") " Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.523570 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83d17d63-8752-4032-a8c1-a0d1770372e8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "83d17d63-8752-4032-a8c1-a0d1770372e8" (UID: "83d17d63-8752-4032-a8c1-a0d1770372e8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.532152 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "83d17d63-8752-4032-a8c1-a0d1770372e8" (UID: "83d17d63-8752-4032-a8c1-a0d1770372e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.543683 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83d17d63-8752-4032-a8c1-a0d1770372e8-kube-api-access-nj5tt" (OuterVolumeSpecName: "kube-api-access-nj5tt") pod "83d17d63-8752-4032-a8c1-a0d1770372e8" (UID: "83d17d63-8752-4032-a8c1-a0d1770372e8"). InnerVolumeSpecName "kube-api-access-nj5tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.543780 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-scripts" (OuterVolumeSpecName: "scripts") pod "83d17d63-8752-4032-a8c1-a0d1770372e8" (UID: "83d17d63-8752-4032-a8c1-a0d1770372e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.595085 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83d17d63-8752-4032-a8c1-a0d1770372e8" (UID: "83d17d63-8752-4032-a8c1-a0d1770372e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.627974 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.628022 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.628031 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.628041 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj5tt\" (UniqueName: \"kubernetes.io/projected/83d17d63-8752-4032-a8c1-a0d1770372e8-kube-api-access-nj5tt\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.628051 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83d17d63-8752-4032-a8c1-a0d1770372e8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.632868 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data" (OuterVolumeSpecName: "config-data") pod "83d17d63-8752-4032-a8c1-a0d1770372e8" (UID: "83d17d63-8752-4032-a8c1-a0d1770372e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.686428 4770 generic.go:334] "Generic (PLEG): container finished" podID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerID="c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d" exitCode=0 Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.686486 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83d17d63-8752-4032-a8c1-a0d1770372e8","Type":"ContainerDied","Data":"c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d"} Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.686513 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83d17d63-8752-4032-a8c1-a0d1770372e8","Type":"ContainerDied","Data":"23acfbe2d593914a91300046a9cca755f0b62c2b21688010078e392feb88e500"} Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.686535 4770 scope.go:117] "RemoveContainer" containerID="6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.686682 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.715897 4770 scope.go:117] "RemoveContainer" containerID="c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.724464 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.730110 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d17d63-8752-4032-a8c1-a0d1770372e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.736270 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.741702 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:22:30 crc kubenswrapper[4770]: E1004 03:22:30.742041 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerName="probe" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.742059 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerName="probe" Oct 04 03:22:30 crc kubenswrapper[4770]: E1004 03:22:30.742089 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerName="cinder-scheduler" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.742095 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerName="cinder-scheduler" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.742284 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerName="cinder-scheduler" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.742308 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d17d63-8752-4032-a8c1-a0d1770372e8" containerName="probe" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.743277 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.751345 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.751843 4770 scope.go:117] "RemoveContainer" containerID="6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.752031 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 03:22:30 crc kubenswrapper[4770]: E1004 03:22:30.756275 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075\": container with ID starting with 6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075 not found: ID does not exist" containerID="6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.756327 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075"} err="failed to get container status \"6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075\": rpc error: code = NotFound desc = could not find container \"6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075\": container with ID starting with 6d33a576744bbf89b63b186f47e0e7de5aacfb7099c7c542249292f72cd5a075 not found: ID does not exist" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.756358 4770 scope.go:117] "RemoveContainer" containerID="c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d" Oct 04 03:22:30 crc kubenswrapper[4770]: E1004 03:22:30.756721 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d\": container with ID starting with c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d not found: ID does not exist" containerID="c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.756742 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d"} err="failed to get container status \"c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d\": rpc error: code = NotFound desc = could not find container \"c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d\": container with ID starting with c4ec2e69d8c9c0b64eb6de30b2d15144e97d07a35b96c6e96d6179fece9bfe1d not found: ID does not exist" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.834951 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.835050 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-scripts\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.835090 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77ql8\" (UniqueName: \"kubernetes.io/projected/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-kube-api-access-77ql8\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.835115 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.835141 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.835174 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.877977 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-zfjdv"] Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.938107 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.938202 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-scripts\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.938250 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77ql8\" (UniqueName: \"kubernetes.io/projected/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-kube-api-access-77ql8\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.938286 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.938337 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.938379 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.938468 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.948900 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-scripts\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.951753 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.954135 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.958906 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.969635 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77ql8\" (UniqueName: \"kubernetes.io/projected/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-kube-api-access-77ql8\") pod \"cinder-scheduler-0\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " pod="openstack/cinder-scheduler-0" Oct 04 03:22:30 crc kubenswrapper[4770]: I1004 03:22:30.997436 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.056963 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f69d8558d-mtg8s"] Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.057199 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f69d8558d-mtg8s" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerName="barbican-api-log" containerID="cri-o://5db9161c1905e1df316b626cd2b33a4cb532dabb24d808c7b021a0e22baa50ef" gracePeriod=30 Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.057570 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f69d8558d-mtg8s" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerName="barbican-api" containerID="cri-o://02071bc128c4c599cd4169231a0d430b639afeec180d3c1bb5d18d6093bf2089" gracePeriod=30 Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.088388 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.089739 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5dbdffbfb6-pnbvk"] Oct 04 03:22:31 crc kubenswrapper[4770]: W1004 03:22:31.111245 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43a28d77_6122_4774_b40d_f92dfdc9933d.slice/crio-bc0c63b76f969dcd85bf1b84c4e7bf1821bdbe5c062f0a685fe453e0e0b37185 WatchSource:0}: Error finding container bc0c63b76f969dcd85bf1b84c4e7bf1821bdbe5c062f0a685fe453e0e0b37185: Status 404 returned error can't find the container with id bc0c63b76f969dcd85bf1b84c4e7bf1821bdbe5c062f0a685fe453e0e0b37185 Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.621717 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:22:31 crc kubenswrapper[4770]: W1004 03:22:31.627611 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1908f1eb_8656_43ee_8532_8b9e4ca8e1c2.slice/crio-e2f9ae58de1810066e718f156dc56e2193540ce888fb2110b94141f99335b8ec WatchSource:0}: Error finding container e2f9ae58de1810066e718f156dc56e2193540ce888fb2110b94141f99335b8ec: Status 404 returned error can't find the container with id e2f9ae58de1810066e718f156dc56e2193540ce888fb2110b94141f99335b8ec Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.693806 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13c7863f-832e-4a86-8ce6-f57c92e834c6" path="/var/lib/kubelet/pods/13c7863f-832e-4a86-8ce6-f57c92e834c6/volumes" Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.694758 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83d17d63-8752-4032-a8c1-a0d1770372e8" path="/var/lib/kubelet/pods/83d17d63-8752-4032-a8c1-a0d1770372e8/volumes" Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.718729 4770 generic.go:334] "Generic (PLEG): container finished" podID="3ac2690b-33cd-48f3-92d9-b9a0e29442e0" containerID="d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d" exitCode=0 Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.719231 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" event={"ID":"3ac2690b-33cd-48f3-92d9-b9a0e29442e0","Type":"ContainerDied","Data":"d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d"} Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.719328 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" event={"ID":"3ac2690b-33cd-48f3-92d9-b9a0e29442e0","Type":"ContainerStarted","Data":"e5234fe27ef714c765747f10c0313ae31c7c9a2d6045523aa1dad71830928543"} Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.722894 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2","Type":"ContainerStarted","Data":"e2f9ae58de1810066e718f156dc56e2193540ce888fb2110b94141f99335b8ec"} Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.726105 4770 generic.go:334] "Generic (PLEG): container finished" podID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerID="5db9161c1905e1df316b626cd2b33a4cb532dabb24d808c7b021a0e22baa50ef" exitCode=143 Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.726259 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69d8558d-mtg8s" event={"ID":"21e58a5a-78a0-4e87-b3ff-5d789c437dd1","Type":"ContainerDied","Data":"5db9161c1905e1df316b626cd2b33a4cb532dabb24d808c7b021a0e22baa50ef"} Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.727482 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dbdffbfb6-pnbvk" event={"ID":"43a28d77-6122-4774-b40d-f92dfdc9933d","Type":"ContainerStarted","Data":"e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d"} Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.727579 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dbdffbfb6-pnbvk" event={"ID":"43a28d77-6122-4774-b40d-f92dfdc9933d","Type":"ContainerStarted","Data":"456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75"} Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.727638 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dbdffbfb6-pnbvk" event={"ID":"43a28d77-6122-4774-b40d-f92dfdc9933d","Type":"ContainerStarted","Data":"bc0c63b76f969dcd85bf1b84c4e7bf1821bdbe5c062f0a685fe453e0e0b37185"} Oct 04 03:22:31 crc kubenswrapper[4770]: I1004 03:22:31.728414 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:22:32 crc kubenswrapper[4770]: I1004 03:22:32.754447 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" event={"ID":"3ac2690b-33cd-48f3-92d9-b9a0e29442e0","Type":"ContainerStarted","Data":"937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd"} Oct 04 03:22:32 crc kubenswrapper[4770]: I1004 03:22:32.755618 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:32 crc kubenswrapper[4770]: I1004 03:22:32.762548 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2","Type":"ContainerStarted","Data":"24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b"} Oct 04 03:22:32 crc kubenswrapper[4770]: I1004 03:22:32.777102 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" podStartSLOduration=3.777085423 podStartE2EDuration="3.777085423s" podCreationTimestamp="2025-10-04 03:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:32.773141579 +0000 UTC m=+1164.065151291" watchObservedRunningTime="2025-10-04 03:22:32.777085423 +0000 UTC m=+1164.069095125" Oct 04 03:22:32 crc kubenswrapper[4770]: I1004 03:22:32.780591 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5dbdffbfb6-pnbvk" podStartSLOduration=2.780581614 podStartE2EDuration="2.780581614s" podCreationTimestamp="2025-10-04 03:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:31.778681712 +0000 UTC m=+1163.070691424" watchObservedRunningTime="2025-10-04 03:22:32.780581614 +0000 UTC m=+1164.072591326" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.248897 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-55cdcb545c-rfq2m"] Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.251679 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.257963 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.258538 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.264777 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55cdcb545c-rfq2m"] Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.405040 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csxq4\" (UniqueName: \"kubernetes.io/projected/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-kube-api-access-csxq4\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.405092 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-combined-ca-bundle\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.405115 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-internal-tls-certs\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.405279 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-ovndb-tls-certs\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.405397 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-config\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.405586 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-public-tls-certs\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.405698 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-httpd-config\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.436735 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.507127 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-httpd-config\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.508319 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csxq4\" (UniqueName: \"kubernetes.io/projected/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-kube-api-access-csxq4\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.508431 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-combined-ca-bundle\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.508508 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-internal-tls-certs\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.508626 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-ovndb-tls-certs\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.508769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-config\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.508886 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-public-tls-certs\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.515596 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-combined-ca-bundle\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.516915 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-ovndb-tls-certs\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.517133 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-httpd-config\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.517696 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-public-tls-certs\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.517802 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-config\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.521411 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-internal-tls-certs\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.528558 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csxq4\" (UniqueName: \"kubernetes.io/projected/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-kube-api-access-csxq4\") pod \"neutron-55cdcb545c-rfq2m\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.573438 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.779986 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2","Type":"ContainerStarted","Data":"1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522"} Oct 04 03:22:33 crc kubenswrapper[4770]: I1004 03:22:33.815899 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.81588192 podStartE2EDuration="3.81588192s" podCreationTimestamp="2025-10-04 03:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:33.815280144 +0000 UTC m=+1165.107289866" watchObservedRunningTime="2025-10-04 03:22:33.81588192 +0000 UTC m=+1165.107891632" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.215554 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.218776 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.220474 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.221692 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.222126 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zrzgk" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.225995 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.327163 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.327512 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppvfm\" (UniqueName: \"kubernetes.io/projected/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-kube-api-access-ppvfm\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.327538 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.327685 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config-secret\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.334040 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55cdcb545c-rfq2m"] Oct 04 03:22:34 crc kubenswrapper[4770]: W1004 03:22:34.354252 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod145bb94b_ff5d_4b2f_97b1_1b225fc4f164.slice/crio-b4f035e89e59b0a3d2e150ddf9a3f79cd9dce5507edc4a105d60c8dcbcfa2951 WatchSource:0}: Error finding container b4f035e89e59b0a3d2e150ddf9a3f79cd9dce5507edc4a105d60c8dcbcfa2951: Status 404 returned error can't find the container with id b4f035e89e59b0a3d2e150ddf9a3f79cd9dce5507edc4a105d60c8dcbcfa2951 Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.429044 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config-secret\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.429122 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.429164 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppvfm\" (UniqueName: \"kubernetes.io/projected/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-kube-api-access-ppvfm\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.429190 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.430239 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.435776 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.448133 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config-secret\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.454711 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppvfm\" (UniqueName: \"kubernetes.io/projected/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-kube-api-access-ppvfm\") pod \"openstackclient\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.537240 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.805189 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55cdcb545c-rfq2m" event={"ID":"145bb94b-ff5d-4b2f-97b1-1b225fc4f164","Type":"ContainerStarted","Data":"b4f035e89e59b0a3d2e150ddf9a3f79cd9dce5507edc4a105d60c8dcbcfa2951"} Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.820900 4770 generic.go:334] "Generic (PLEG): container finished" podID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerID="02071bc128c4c599cd4169231a0d430b639afeec180d3c1bb5d18d6093bf2089" exitCode=0 Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.821859 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69d8558d-mtg8s" event={"ID":"21e58a5a-78a0-4e87-b3ff-5d789c437dd1","Type":"ContainerDied","Data":"02071bc128c4c599cd4169231a0d430b639afeec180d3c1bb5d18d6093bf2089"} Oct 04 03:22:34 crc kubenswrapper[4770]: I1004 03:22:34.940293 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.046449 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-combined-ca-bundle\") pod \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.046831 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data\") pod \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.046961 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data-custom\") pod \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.046980 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf6vh\" (UniqueName: \"kubernetes.io/projected/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-kube-api-access-rf6vh\") pod \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.047017 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-logs\") pod \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\" (UID: \"21e58a5a-78a0-4e87-b3ff-5d789c437dd1\") " Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.048415 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-logs" (OuterVolumeSpecName: "logs") pod "21e58a5a-78a0-4e87-b3ff-5d789c437dd1" (UID: "21e58a5a-78a0-4e87-b3ff-5d789c437dd1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.054127 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "21e58a5a-78a0-4e87-b3ff-5d789c437dd1" (UID: "21e58a5a-78a0-4e87-b3ff-5d789c437dd1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.059596 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-kube-api-access-rf6vh" (OuterVolumeSpecName: "kube-api-access-rf6vh") pod "21e58a5a-78a0-4e87-b3ff-5d789c437dd1" (UID: "21e58a5a-78a0-4e87-b3ff-5d789c437dd1"). InnerVolumeSpecName "kube-api-access-rf6vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.115803 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21e58a5a-78a0-4e87-b3ff-5d789c437dd1" (UID: "21e58a5a-78a0-4e87-b3ff-5d789c437dd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.116106 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data" (OuterVolumeSpecName: "config-data") pod "21e58a5a-78a0-4e87-b3ff-5d789c437dd1" (UID: "21e58a5a-78a0-4e87-b3ff-5d789c437dd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.151352 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.151493 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf6vh\" (UniqueName: \"kubernetes.io/projected/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-kube-api-access-rf6vh\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.151508 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.151521 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.151529 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e58a5a-78a0-4e87-b3ff-5d789c437dd1-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.178364 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.837899 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55cdcb545c-rfq2m" event={"ID":"145bb94b-ff5d-4b2f-97b1-1b225fc4f164","Type":"ContainerStarted","Data":"886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97"} Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.838484 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55cdcb545c-rfq2m" event={"ID":"145bb94b-ff5d-4b2f-97b1-1b225fc4f164","Type":"ContainerStarted","Data":"fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85"} Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.838943 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.840857 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5d67ab7f-c0b5-476c-b5d2-99b90eed466a","Type":"ContainerStarted","Data":"c20e0a66e20732fc703aa106bc5fb69a0775739552466a7ea981128f4861080f"} Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.844259 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69d8558d-mtg8s" event={"ID":"21e58a5a-78a0-4e87-b3ff-5d789c437dd1","Type":"ContainerDied","Data":"7d918b4e978c7c4ffe7dc40f7ce0ad21b0971dddf7167f02d2d5f1fb3c05af9f"} Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.844301 4770 scope.go:117] "RemoveContainer" containerID="02071bc128c4c599cd4169231a0d430b639afeec180d3c1bb5d18d6093bf2089" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.844577 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f69d8558d-mtg8s" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.864155 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-55cdcb545c-rfq2m" podStartSLOduration=2.864135016 podStartE2EDuration="2.864135016s" podCreationTimestamp="2025-10-04 03:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:35.863193172 +0000 UTC m=+1167.155202894" watchObservedRunningTime="2025-10-04 03:22:35.864135016 +0000 UTC m=+1167.156144728" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.886621 4770 scope.go:117] "RemoveContainer" containerID="5db9161c1905e1df316b626cd2b33a4cb532dabb24d808c7b021a0e22baa50ef" Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.888893 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f69d8558d-mtg8s"] Oct 04 03:22:35 crc kubenswrapper[4770]: I1004 03:22:35.908465 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6f69d8558d-mtg8s"] Oct 04 03:22:36 crc kubenswrapper[4770]: I1004 03:22:36.089798 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 03:22:36 crc kubenswrapper[4770]: I1004 03:22:36.454838 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 03:22:37 crc kubenswrapper[4770]: I1004 03:22:37.689859 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" path="/var/lib/kubelet/pods/21e58a5a-78a0-4e87-b3ff-5d789c437dd1/volumes" Oct 04 03:22:40 crc kubenswrapper[4770]: I1004 03:22:40.276888 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:22:40 crc kubenswrapper[4770]: I1004 03:22:40.332398 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7889fffdf7-snw87"] Oct 04 03:22:40 crc kubenswrapper[4770]: I1004 03:22:40.332630 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" podUID="ba829814-c679-46a0-99ed-d7a04e2bf96d" containerName="dnsmasq-dns" containerID="cri-o://cdee7b6e2ac49a189d7a69bac5314a7d4296f6bd3b020ac64a10b6275ad7a4f6" gracePeriod=10 Oct 04 03:22:40 crc kubenswrapper[4770]: I1004 03:22:40.911968 4770 generic.go:334] "Generic (PLEG): container finished" podID="ba829814-c679-46a0-99ed-d7a04e2bf96d" containerID="cdee7b6e2ac49a189d7a69bac5314a7d4296f6bd3b020ac64a10b6275ad7a4f6" exitCode=0 Oct 04 03:22:40 crc kubenswrapper[4770]: I1004 03:22:40.912157 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" event={"ID":"ba829814-c679-46a0-99ed-d7a04e2bf96d","Type":"ContainerDied","Data":"cdee7b6e2ac49a189d7a69bac5314a7d4296f6bd3b020ac64a10b6275ad7a4f6"} Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.324618 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-787bf4d777-x8gln"] Oct 04 03:22:41 crc kubenswrapper[4770]: E1004 03:22:41.325392 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerName="barbican-api-log" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.325421 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerName="barbican-api-log" Oct 04 03:22:41 crc kubenswrapper[4770]: E1004 03:22:41.325439 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerName="barbican-api" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.325445 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerName="barbican-api" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.325669 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerName="barbican-api" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.325690 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e58a5a-78a0-4e87-b3ff-5d789c437dd1" containerName="barbican-api-log" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.326877 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.331266 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.331458 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.331528 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.340747 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-787bf4d777-x8gln"] Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.370339 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.417997 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-log-httpd\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.418073 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-etc-swift\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.418103 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-run-httpd\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.418150 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-config-data\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.418187 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-internal-tls-certs\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.418219 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx8hd\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-kube-api-access-rx8hd\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.418241 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-combined-ca-bundle\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.418263 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-public-tls-certs\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.519902 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-combined-ca-bundle\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.519958 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-public-tls-certs\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.520126 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-log-httpd\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.520175 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-etc-swift\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.520194 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-run-httpd\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.520255 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-config-data\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.520296 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-internal-tls-certs\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.520326 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx8hd\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-kube-api-access-rx8hd\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.521206 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-run-httpd\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.522375 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-log-httpd\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.527706 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-config-data\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.528449 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-public-tls-certs\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.534687 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-combined-ca-bundle\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.535293 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-internal-tls-certs\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.537401 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-etc-swift\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.543366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx8hd\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-kube-api-access-rx8hd\") pod \"swift-proxy-787bf4d777-x8gln\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:41 crc kubenswrapper[4770]: I1004 03:22:41.648283 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.395525 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.396218 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d4377592-806c-4c05-9527-fe09c7802b73" containerName="glance-log" containerID="cri-o://a0f9b2c44a0420584b66e0af2a755d11a35c8d018901f1ba39dec7459258e5a3" gracePeriod=30 Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.396357 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d4377592-806c-4c05-9527-fe09c7802b73" containerName="glance-httpd" containerID="cri-o://28014651bae2a240ed227b651ddc14ed3143c2ea8cc47d1b44c7a6b98514b480" gracePeriod=30 Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.407414 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="d4377592-806c-4c05-9527-fe09c7802b73" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.146:9292/healthcheck\": EOF" Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.803438 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.803997 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="ceilometer-central-agent" containerID="cri-o://c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c" gracePeriod=30 Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.804077 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="sg-core" containerID="cri-o://a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab" gracePeriod=30 Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.804095 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="proxy-httpd" containerID="cri-o://5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd" gracePeriod=30 Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.804132 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="ceilometer-notification-agent" containerID="cri-o://7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753" gracePeriod=30 Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.819221 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.949893 4770 generic.go:334] "Generic (PLEG): container finished" podID="d4377592-806c-4c05-9527-fe09c7802b73" containerID="a0f9b2c44a0420584b66e0af2a755d11a35c8d018901f1ba39dec7459258e5a3" exitCode=143 Oct 04 03:22:42 crc kubenswrapper[4770]: I1004 03:22:42.949980 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d4377592-806c-4c05-9527-fe09c7802b73","Type":"ContainerDied","Data":"a0f9b2c44a0420584b66e0af2a755d11a35c8d018901f1ba39dec7459258e5a3"} Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.074761 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" podUID="ba829814-c679-46a0-99ed-d7a04e2bf96d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.158:5353: connect: connection refused" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.151458 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-n8mcv"] Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.153026 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-n8mcv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.172581 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-n8mcv"] Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.249845 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-pv8zj"] Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.251088 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-pv8zj" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.254576 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czxjf\" (UniqueName: \"kubernetes.io/projected/3c9fa8cd-207f-4561-9116-ceea539098d8-kube-api-access-czxjf\") pod \"nova-api-db-create-n8mcv\" (UID: \"3c9fa8cd-207f-4561-9116-ceea539098d8\") " pod="openstack/nova-api-db-create-n8mcv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.262090 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-pv8zj"] Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.351977 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-kv9tv"] Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.353196 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kv9tv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.357272 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrhrf\" (UniqueName: \"kubernetes.io/projected/cbd7579a-ef76-48fc-96f6-3b26466084f8-kube-api-access-hrhrf\") pod \"nova-cell0-db-create-pv8zj\" (UID: \"cbd7579a-ef76-48fc-96f6-3b26466084f8\") " pod="openstack/nova-cell0-db-create-pv8zj" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.357401 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czxjf\" (UniqueName: \"kubernetes.io/projected/3c9fa8cd-207f-4561-9116-ceea539098d8-kube-api-access-czxjf\") pod \"nova-api-db-create-n8mcv\" (UID: \"3c9fa8cd-207f-4561-9116-ceea539098d8\") " pod="openstack/nova-api-db-create-n8mcv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.373943 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kv9tv"] Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.383222 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czxjf\" (UniqueName: \"kubernetes.io/projected/3c9fa8cd-207f-4561-9116-ceea539098d8-kube-api-access-czxjf\") pod \"nova-api-db-create-n8mcv\" (UID: \"3c9fa8cd-207f-4561-9116-ceea539098d8\") " pod="openstack/nova-api-db-create-n8mcv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.459515 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px789\" (UniqueName: \"kubernetes.io/projected/23bf2c46-426f-4bb7-82e0-f9e90d0a2f32-kube-api-access-px789\") pod \"nova-cell1-db-create-kv9tv\" (UID: \"23bf2c46-426f-4bb7-82e0-f9e90d0a2f32\") " pod="openstack/nova-cell1-db-create-kv9tv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.459584 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrhrf\" (UniqueName: \"kubernetes.io/projected/cbd7579a-ef76-48fc-96f6-3b26466084f8-kube-api-access-hrhrf\") pod \"nova-cell0-db-create-pv8zj\" (UID: \"cbd7579a-ef76-48fc-96f6-3b26466084f8\") " pod="openstack/nova-cell0-db-create-pv8zj" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.480144 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrhrf\" (UniqueName: \"kubernetes.io/projected/cbd7579a-ef76-48fc-96f6-3b26466084f8-kube-api-access-hrhrf\") pod \"nova-cell0-db-create-pv8zj\" (UID: \"cbd7579a-ef76-48fc-96f6-3b26466084f8\") " pod="openstack/nova-cell0-db-create-pv8zj" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.482059 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-n8mcv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.560892 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px789\" (UniqueName: \"kubernetes.io/projected/23bf2c46-426f-4bb7-82e0-f9e90d0a2f32-kube-api-access-px789\") pod \"nova-cell1-db-create-kv9tv\" (UID: \"23bf2c46-426f-4bb7-82e0-f9e90d0a2f32\") " pod="openstack/nova-cell1-db-create-kv9tv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.581045 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-pv8zj" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.583586 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px789\" (UniqueName: \"kubernetes.io/projected/23bf2c46-426f-4bb7-82e0-f9e90d0a2f32-kube-api-access-px789\") pod \"nova-cell1-db-create-kv9tv\" (UID: \"23bf2c46-426f-4bb7-82e0-f9e90d0a2f32\") " pod="openstack/nova-cell1-db-create-kv9tv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.677564 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kv9tv" Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.964143 4770 generic.go:334] "Generic (PLEG): container finished" podID="3626db89-39a8-40b8-ae68-e900090e76e9" containerID="5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd" exitCode=0 Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.964189 4770 generic.go:334] "Generic (PLEG): container finished" podID="3626db89-39a8-40b8-ae68-e900090e76e9" containerID="a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab" exitCode=2 Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.964201 4770 generic.go:334] "Generic (PLEG): container finished" podID="3626db89-39a8-40b8-ae68-e900090e76e9" containerID="c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c" exitCode=0 Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.964220 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerDied","Data":"5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd"} Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.964265 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerDied","Data":"a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab"} Oct 04 03:22:43 crc kubenswrapper[4770]: I1004 03:22:43.964277 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerDied","Data":"c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c"} Oct 04 03:22:46 crc kubenswrapper[4770]: I1004 03:22:46.586434 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.156:3000/\": dial tcp 10.217.0.156:3000: connect: connection refused" Oct 04 03:22:46 crc kubenswrapper[4770]: I1004 03:22:46.784714 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:22:46 crc kubenswrapper[4770]: I1004 03:22:46.784968 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="94883722-c05e-43f6-ac7d-91ce20de293b" containerName="glance-log" containerID="cri-o://36527a07ae6792b10bdaf728b57e11696b53baf86da649a634ef2e860974b76b" gracePeriod=30 Oct 04 03:22:46 crc kubenswrapper[4770]: I1004 03:22:46.785142 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="94883722-c05e-43f6-ac7d-91ce20de293b" containerName="glance-httpd" containerID="cri-o://9d13162cc33717d452f99db8ca695a367a80b9d5bc0d24acc6684388fef67953" gracePeriod=30 Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.007142 4770 generic.go:334] "Generic (PLEG): container finished" podID="94883722-c05e-43f6-ac7d-91ce20de293b" containerID="36527a07ae6792b10bdaf728b57e11696b53baf86da649a634ef2e860974b76b" exitCode=143 Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.007189 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94883722-c05e-43f6-ac7d-91ce20de293b","Type":"ContainerDied","Data":"36527a07ae6792b10bdaf728b57e11696b53baf86da649a634ef2e860974b76b"} Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.011599 4770 generic.go:334] "Generic (PLEG): container finished" podID="d4377592-806c-4c05-9527-fe09c7802b73" containerID="28014651bae2a240ed227b651ddc14ed3143c2ea8cc47d1b44c7a6b98514b480" exitCode=0 Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.011640 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d4377592-806c-4c05-9527-fe09c7802b73","Type":"ContainerDied","Data":"28014651bae2a240ed227b651ddc14ed3143c2ea8cc47d1b44c7a6b98514b480"} Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.366480 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.544868 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-config\") pod \"ba829814-c679-46a0-99ed-d7a04e2bf96d\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.545029 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-nb\") pod \"ba829814-c679-46a0-99ed-d7a04e2bf96d\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.545086 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-svc\") pod \"ba829814-c679-46a0-99ed-d7a04e2bf96d\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.545141 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-swift-storage-0\") pod \"ba829814-c679-46a0-99ed-d7a04e2bf96d\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.545178 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-sb\") pod \"ba829814-c679-46a0-99ed-d7a04e2bf96d\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.546623 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prrwq\" (UniqueName: \"kubernetes.io/projected/ba829814-c679-46a0-99ed-d7a04e2bf96d-kube-api-access-prrwq\") pod \"ba829814-c679-46a0-99ed-d7a04e2bf96d\" (UID: \"ba829814-c679-46a0-99ed-d7a04e2bf96d\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.557999 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba829814-c679-46a0-99ed-d7a04e2bf96d-kube-api-access-prrwq" (OuterVolumeSpecName: "kube-api-access-prrwq") pod "ba829814-c679-46a0-99ed-d7a04e2bf96d" (UID: "ba829814-c679-46a0-99ed-d7a04e2bf96d"). InnerVolumeSpecName "kube-api-access-prrwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.603911 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ba829814-c679-46a0-99ed-d7a04e2bf96d" (UID: "ba829814-c679-46a0-99ed-d7a04e2bf96d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.603952 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ba829814-c679-46a0-99ed-d7a04e2bf96d" (UID: "ba829814-c679-46a0-99ed-d7a04e2bf96d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.609928 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.615350 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ba829814-c679-46a0-99ed-d7a04e2bf96d" (UID: "ba829814-c679-46a0-99ed-d7a04e2bf96d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.621078 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ba829814-c679-46a0-99ed-d7a04e2bf96d" (UID: "ba829814-c679-46a0-99ed-d7a04e2bf96d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.622348 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-config" (OuterVolumeSpecName: "config") pod "ba829814-c679-46a0-99ed-d7a04e2bf96d" (UID: "ba829814-c679-46a0-99ed-d7a04e2bf96d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.651741 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.651785 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.651796 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.651805 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.651814 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prrwq\" (UniqueName: \"kubernetes.io/projected/ba829814-c679-46a0-99ed-d7a04e2bf96d-kube-api-access-prrwq\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.651826 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba829814-c679-46a0-99ed-d7a04e2bf96d-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.760195 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-logs\") pod \"d4377592-806c-4c05-9527-fe09c7802b73\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.761786 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-combined-ca-bundle\") pod \"d4377592-806c-4c05-9527-fe09c7802b73\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.761965 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-scripts\") pod \"d4377592-806c-4c05-9527-fe09c7802b73\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.762175 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhsl6\" (UniqueName: \"kubernetes.io/projected/d4377592-806c-4c05-9527-fe09c7802b73-kube-api-access-zhsl6\") pod \"d4377592-806c-4c05-9527-fe09c7802b73\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.762346 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d4377592-806c-4c05-9527-fe09c7802b73\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.762452 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-httpd-run\") pod \"d4377592-806c-4c05-9527-fe09c7802b73\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.762659 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-public-tls-certs\") pod \"d4377592-806c-4c05-9527-fe09c7802b73\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.762809 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-config-data\") pod \"d4377592-806c-4c05-9527-fe09c7802b73\" (UID: \"d4377592-806c-4c05-9527-fe09c7802b73\") " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.767347 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kv9tv"] Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.760783 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-logs" (OuterVolumeSpecName: "logs") pod "d4377592-806c-4c05-9527-fe09c7802b73" (UID: "d4377592-806c-4c05-9527-fe09c7802b73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.772484 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "d4377592-806c-4c05-9527-fe09c7802b73" (UID: "d4377592-806c-4c05-9527-fe09c7802b73"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.773143 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4377592-806c-4c05-9527-fe09c7802b73-kube-api-access-zhsl6" (OuterVolumeSpecName: "kube-api-access-zhsl6") pod "d4377592-806c-4c05-9527-fe09c7802b73" (UID: "d4377592-806c-4c05-9527-fe09c7802b73"). InnerVolumeSpecName "kube-api-access-zhsl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.773845 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d4377592-806c-4c05-9527-fe09c7802b73" (UID: "d4377592-806c-4c05-9527-fe09c7802b73"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.781263 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-scripts" (OuterVolumeSpecName: "scripts") pod "d4377592-806c-4c05-9527-fe09c7802b73" (UID: "d4377592-806c-4c05-9527-fe09c7802b73"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.781271 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-n8mcv"] Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.806585 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4377592-806c-4c05-9527-fe09c7802b73" (UID: "d4377592-806c-4c05-9527-fe09c7802b73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.849167 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-config-data" (OuterVolumeSpecName: "config-data") pod "d4377592-806c-4c05-9527-fe09c7802b73" (UID: "d4377592-806c-4c05-9527-fe09c7802b73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.865181 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.865772 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.865798 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.865810 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.865819 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhsl6\" (UniqueName: \"kubernetes.io/projected/d4377592-806c-4c05-9527-fe09c7802b73-kube-api-access-zhsl6\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.865839 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.865848 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4377592-806c-4c05-9527-fe09c7802b73-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.876917 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d4377592-806c-4c05-9527-fe09c7802b73" (UID: "d4377592-806c-4c05-9527-fe09c7802b73"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.900984 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.967235 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.967279 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4377592-806c-4c05-9527-fe09c7802b73-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:47 crc kubenswrapper[4770]: I1004 03:22:47.972164 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-pv8zj"] Oct 04 03:22:47 crc kubenswrapper[4770]: W1004 03:22:47.985620 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbd7579a_ef76_48fc_96f6_3b26466084f8.slice/crio-687f46dfbfb423c36674fba2ad86d4252644374bb8a33659bd602108047a4520 WatchSource:0}: Error finding container 687f46dfbfb423c36674fba2ad86d4252644374bb8a33659bd602108047a4520: Status 404 returned error can't find the container with id 687f46dfbfb423c36674fba2ad86d4252644374bb8a33659bd602108047a4520 Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.025104 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5d67ab7f-c0b5-476c-b5d2-99b90eed466a","Type":"ContainerStarted","Data":"f4087f933afeeb3b260782f8334f21234028523075122f703a71f7c541bb19dc"} Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.028600 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-n8mcv" event={"ID":"3c9fa8cd-207f-4561-9116-ceea539098d8","Type":"ContainerStarted","Data":"58bfdd93d512adbe87b4be0e04e74adfee679c60d791c8c3e70c035d7608b780"} Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.030757 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" event={"ID":"ba829814-c679-46a0-99ed-d7a04e2bf96d","Type":"ContainerDied","Data":"b170c0fe1a38cb16c9349d907191b2b9febc4da2bcdd964a791e280c71fe7de0"} Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.030806 4770 scope.go:117] "RemoveContainer" containerID="cdee7b6e2ac49a189d7a69bac5314a7d4296f6bd3b020ac64a10b6275ad7a4f6" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.030921 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7889fffdf7-snw87" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.033497 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-pv8zj" event={"ID":"cbd7579a-ef76-48fc-96f6-3b26466084f8","Type":"ContainerStarted","Data":"687f46dfbfb423c36674fba2ad86d4252644374bb8a33659bd602108047a4520"} Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.036593 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kv9tv" event={"ID":"23bf2c46-426f-4bb7-82e0-f9e90d0a2f32","Type":"ContainerStarted","Data":"59d4338ba8f0336c488c1c0fd551f96cb38777c8f9ffde34fb2b9cf9a7b1889c"} Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.044619 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d4377592-806c-4c05-9527-fe09c7802b73","Type":"ContainerDied","Data":"2c2d6b0f430cc36ff0f138569977ab3ba444759f49ad43944b5906f5b23f8350"} Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.044746 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.048918 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.173717636 podStartE2EDuration="14.048895268s" podCreationTimestamp="2025-10-04 03:22:34 +0000 UTC" firstStartedPulling="2025-10-04 03:22:35.177978675 +0000 UTC m=+1166.469988387" lastFinishedPulling="2025-10-04 03:22:47.053156307 +0000 UTC m=+1178.345166019" observedRunningTime="2025-10-04 03:22:48.039100662 +0000 UTC m=+1179.331110374" watchObservedRunningTime="2025-10-04 03:22:48.048895268 +0000 UTC m=+1179.340904980" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.149430 4770 scope.go:117] "RemoveContainer" containerID="80067f66c1954d7df70f8381839b5e771c4dde9379cad0ef31951faee78c84e0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.244633 4770 scope.go:117] "RemoveContainer" containerID="28014651bae2a240ed227b651ddc14ed3143c2ea8cc47d1b44c7a6b98514b480" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.259070 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7889fffdf7-snw87"] Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.290794 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7889fffdf7-snw87"] Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.322523 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.328500 4770 scope.go:117] "RemoveContainer" containerID="a0f9b2c44a0420584b66e0af2a755d11a35c8d018901f1ba39dec7459258e5a3" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.344081 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.354143 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:22:48 crc kubenswrapper[4770]: E1004 03:22:48.354792 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4377592-806c-4c05-9527-fe09c7802b73" containerName="glance-httpd" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.354812 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4377592-806c-4c05-9527-fe09c7802b73" containerName="glance-httpd" Oct 04 03:22:48 crc kubenswrapper[4770]: E1004 03:22:48.354821 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba829814-c679-46a0-99ed-d7a04e2bf96d" containerName="init" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.354828 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba829814-c679-46a0-99ed-d7a04e2bf96d" containerName="init" Oct 04 03:22:48 crc kubenswrapper[4770]: E1004 03:22:48.354874 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba829814-c679-46a0-99ed-d7a04e2bf96d" containerName="dnsmasq-dns" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.354880 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba829814-c679-46a0-99ed-d7a04e2bf96d" containerName="dnsmasq-dns" Oct 04 03:22:48 crc kubenswrapper[4770]: E1004 03:22:48.354902 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4377592-806c-4c05-9527-fe09c7802b73" containerName="glance-log" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.354907 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4377592-806c-4c05-9527-fe09c7802b73" containerName="glance-log" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.355140 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba829814-c679-46a0-99ed-d7a04e2bf96d" containerName="dnsmasq-dns" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.355157 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4377592-806c-4c05-9527-fe09c7802b73" containerName="glance-httpd" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.355178 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4377592-806c-4c05-9527-fe09c7802b73" containerName="glance-log" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.364638 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.370553 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.371468 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.378466 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.436359 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.499746 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.499798 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-logs\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.499832 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-scripts\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.499861 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9sdf\" (UniqueName: \"kubernetes.io/projected/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-kube-api-access-w9sdf\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.500113 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-config-data\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.500186 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.500298 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.500321 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.601630 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-scripts\") pod \"3626db89-39a8-40b8-ae68-e900090e76e9\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.601708 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-sg-core-conf-yaml\") pod \"3626db89-39a8-40b8-ae68-e900090e76e9\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.601748 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-config-data\") pod \"3626db89-39a8-40b8-ae68-e900090e76e9\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.601835 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-log-httpd\") pod \"3626db89-39a8-40b8-ae68-e900090e76e9\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.601877 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-run-httpd\") pod \"3626db89-39a8-40b8-ae68-e900090e76e9\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.601913 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-combined-ca-bundle\") pod \"3626db89-39a8-40b8-ae68-e900090e76e9\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.602023 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr55g\" (UniqueName: \"kubernetes.io/projected/3626db89-39a8-40b8-ae68-e900090e76e9-kube-api-access-nr55g\") pod \"3626db89-39a8-40b8-ae68-e900090e76e9\" (UID: \"3626db89-39a8-40b8-ae68-e900090e76e9\") " Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.602253 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-scripts\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.602295 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9sdf\" (UniqueName: \"kubernetes.io/projected/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-kube-api-access-w9sdf\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.602357 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-config-data\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.602388 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.602461 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.602483 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.602541 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.602575 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-logs\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.603112 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-logs\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.603325 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.603423 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.604062 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3626db89-39a8-40b8-ae68-e900090e76e9" (UID: "3626db89-39a8-40b8-ae68-e900090e76e9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.606371 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3626db89-39a8-40b8-ae68-e900090e76e9" (UID: "3626db89-39a8-40b8-ae68-e900090e76e9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.608759 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.612153 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.612172 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-scripts\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.613751 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3626db89-39a8-40b8-ae68-e900090e76e9-kube-api-access-nr55g" (OuterVolumeSpecName: "kube-api-access-nr55g") pod "3626db89-39a8-40b8-ae68-e900090e76e9" (UID: "3626db89-39a8-40b8-ae68-e900090e76e9"). InnerVolumeSpecName "kube-api-access-nr55g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.619676 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-scripts" (OuterVolumeSpecName: "scripts") pod "3626db89-39a8-40b8-ae68-e900090e76e9" (UID: "3626db89-39a8-40b8-ae68-e900090e76e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.621248 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-config-data\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.623432 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9sdf\" (UniqueName: \"kubernetes.io/projected/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-kube-api-access-w9sdf\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.646708 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3626db89-39a8-40b8-ae68-e900090e76e9" (UID: "3626db89-39a8-40b8-ae68-e900090e76e9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.657633 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.705124 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr55g\" (UniqueName: \"kubernetes.io/projected/3626db89-39a8-40b8-ae68-e900090e76e9-kube-api-access-nr55g\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.705163 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.705176 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.705188 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.705203 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3626db89-39a8-40b8-ae68-e900090e76e9-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.711504 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3626db89-39a8-40b8-ae68-e900090e76e9" (UID: "3626db89-39a8-40b8-ae68-e900090e76e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.723262 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.736467 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-config-data" (OuterVolumeSpecName: "config-data") pod "3626db89-39a8-40b8-ae68-e900090e76e9" (UID: "3626db89-39a8-40b8-ae68-e900090e76e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.807538 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:48 crc kubenswrapper[4770]: I1004 03:22:48.808000 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3626db89-39a8-40b8-ae68-e900090e76e9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.017118 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-787bf4d777-x8gln"] Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.058812 4770 generic.go:334] "Generic (PLEG): container finished" podID="3626db89-39a8-40b8-ae68-e900090e76e9" containerID="7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753" exitCode=0 Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.059867 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.059808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerDied","Data":"7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753"} Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.059954 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3626db89-39a8-40b8-ae68-e900090e76e9","Type":"ContainerDied","Data":"aec713cc7199ab192a6b063d2ec83799e2c6f6d42c114a64457468efbbfd25cf"} Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.059981 4770 scope.go:117] "RemoveContainer" containerID="5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.066321 4770 generic.go:334] "Generic (PLEG): container finished" podID="cbd7579a-ef76-48fc-96f6-3b26466084f8" containerID="999a5b5b19028ec70d3c4019c51d43f906d5ba05dfa731441b8d79c2228d7297" exitCode=0 Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.066446 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-pv8zj" event={"ID":"cbd7579a-ef76-48fc-96f6-3b26466084f8","Type":"ContainerDied","Data":"999a5b5b19028ec70d3c4019c51d43f906d5ba05dfa731441b8d79c2228d7297"} Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.072490 4770 generic.go:334] "Generic (PLEG): container finished" podID="23bf2c46-426f-4bb7-82e0-f9e90d0a2f32" containerID="2a04b73716f5487a1c885d411e35e23f8c38353fc1a171192c915f57244bfe0d" exitCode=0 Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.072602 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kv9tv" event={"ID":"23bf2c46-426f-4bb7-82e0-f9e90d0a2f32","Type":"ContainerDied","Data":"2a04b73716f5487a1c885d411e35e23f8c38353fc1a171192c915f57244bfe0d"} Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.079636 4770 generic.go:334] "Generic (PLEG): container finished" podID="3c9fa8cd-207f-4561-9116-ceea539098d8" containerID="f79f250529ae5445c1fa9f8a7d6af7dcafd0d968c4d30fd4ce6ee7ab033b040a" exitCode=0 Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.079720 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-n8mcv" event={"ID":"3c9fa8cd-207f-4561-9116-ceea539098d8","Type":"ContainerDied","Data":"f79f250529ae5445c1fa9f8a7d6af7dcafd0d968c4d30fd4ce6ee7ab033b040a"} Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.084168 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-787bf4d777-x8gln" event={"ID":"06da071b-e1a1-42da-9e6e-fd957923876d","Type":"ContainerStarted","Data":"38c0143552fdfa921f35db23330f1117d2705fefd42165f2e2c2ee0e96be9c27"} Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.232165 4770 scope.go:117] "RemoveContainer" containerID="a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.277216 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.279534 4770 scope.go:117] "RemoveContainer" containerID="7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.295067 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.327411 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:49 crc kubenswrapper[4770]: E1004 03:22:49.327804 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="proxy-httpd" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.327817 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="proxy-httpd" Oct 04 03:22:49 crc kubenswrapper[4770]: E1004 03:22:49.327849 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="ceilometer-central-agent" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.327855 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="ceilometer-central-agent" Oct 04 03:22:49 crc kubenswrapper[4770]: E1004 03:22:49.327868 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="sg-core" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.327874 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="sg-core" Oct 04 03:22:49 crc kubenswrapper[4770]: E1004 03:22:49.327893 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="ceilometer-notification-agent" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.327899 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="ceilometer-notification-agent" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.328066 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="proxy-httpd" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.328083 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="ceilometer-notification-agent" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.328099 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="ceilometer-central-agent" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.328110 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" containerName="sg-core" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.329598 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.335316 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.335801 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.339057 4770 scope.go:117] "RemoveContainer" containerID="c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.357319 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.432356 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-scripts\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.432403 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-config-data\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.432423 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9c2z\" (UniqueName: \"kubernetes.io/projected/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-kube-api-access-j9c2z\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.432471 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-log-httpd\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.432499 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.432522 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.432548 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-run-httpd\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.434574 4770 scope.go:117] "RemoveContainer" containerID="5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd" Oct 04 03:22:49 crc kubenswrapper[4770]: E1004 03:22:49.436585 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd\": container with ID starting with 5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd not found: ID does not exist" containerID="5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.436633 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd"} err="failed to get container status \"5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd\": rpc error: code = NotFound desc = could not find container \"5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd\": container with ID starting with 5ef6600aafc5a3a87d2ca21ad491a3762ab1bb9227e9eb78f3887502ab1e4cdd not found: ID does not exist" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.436663 4770 scope.go:117] "RemoveContainer" containerID="a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab" Oct 04 03:22:49 crc kubenswrapper[4770]: E1004 03:22:49.437087 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab\": container with ID starting with a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab not found: ID does not exist" containerID="a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.437123 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab"} err="failed to get container status \"a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab\": rpc error: code = NotFound desc = could not find container \"a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab\": container with ID starting with a0c2c109f23811e65ec60710fa2421b2c896d66ff9efcf5e3a2f23b966901fab not found: ID does not exist" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.437141 4770 scope.go:117] "RemoveContainer" containerID="7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753" Oct 04 03:22:49 crc kubenswrapper[4770]: E1004 03:22:49.437625 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753\": container with ID starting with 7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753 not found: ID does not exist" containerID="7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.437651 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753"} err="failed to get container status \"7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753\": rpc error: code = NotFound desc = could not find container \"7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753\": container with ID starting with 7b338085226f8ab2f2f37e90a63edd0e4ca3bdc5736de59de7f75563612f8753 not found: ID does not exist" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.437671 4770 scope.go:117] "RemoveContainer" containerID="c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c" Oct 04 03:22:49 crc kubenswrapper[4770]: E1004 03:22:49.437864 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c\": container with ID starting with c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c not found: ID does not exist" containerID="c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.437892 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c"} err="failed to get container status \"c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c\": rpc error: code = NotFound desc = could not find container \"c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c\": container with ID starting with c0721df10cee053c6fb4ac6a864e4cf969f7f51d6337e60b77b3b3d5602d3f1c not found: ID does not exist" Oct 04 03:22:49 crc kubenswrapper[4770]: W1004 03:22:49.484165 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cb81588_b9f4_4d8a_8baf_33aebaa6e64d.slice/crio-ae4f904c53b0cb037ba79bc489ee1fcf211d72c4af3c433201d4c6599fddff11 WatchSource:0}: Error finding container ae4f904c53b0cb037ba79bc489ee1fcf211d72c4af3c433201d4c6599fddff11: Status 404 returned error can't find the container with id ae4f904c53b0cb037ba79bc489ee1fcf211d72c4af3c433201d4c6599fddff11 Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.488758 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.533643 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-scripts\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.533689 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-config-data\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.533709 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9c2z\" (UniqueName: \"kubernetes.io/projected/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-kube-api-access-j9c2z\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.533756 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-log-httpd\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.533787 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.533812 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.533833 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-run-httpd\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.534435 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-run-httpd\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.534853 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-log-httpd\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.537853 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-config-data\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.540050 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.541036 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.541604 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-scripts\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.555942 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9c2z\" (UniqueName: \"kubernetes.io/projected/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-kube-api-access-j9c2z\") pod \"ceilometer-0\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.696731 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.761620 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3626db89-39a8-40b8-ae68-e900090e76e9" path="/var/lib/kubelet/pods/3626db89-39a8-40b8-ae68-e900090e76e9/volumes" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.763022 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba829814-c679-46a0-99ed-d7a04e2bf96d" path="/var/lib/kubelet/pods/ba829814-c679-46a0-99ed-d7a04e2bf96d/volumes" Oct 04 03:22:49 crc kubenswrapper[4770]: I1004 03:22:49.770860 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4377592-806c-4c05-9527-fe09c7802b73" path="/var/lib/kubelet/pods/d4377592-806c-4c05-9527-fe09c7802b73/volumes" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.116403 4770 generic.go:334] "Generic (PLEG): container finished" podID="94883722-c05e-43f6-ac7d-91ce20de293b" containerID="9d13162cc33717d452f99db8ca695a367a80b9d5bc0d24acc6684388fef67953" exitCode=0 Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.116711 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94883722-c05e-43f6-ac7d-91ce20de293b","Type":"ContainerDied","Data":"9d13162cc33717d452f99db8ca695a367a80b9d5bc0d24acc6684388fef67953"} Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.118923 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d","Type":"ContainerStarted","Data":"ae4f904c53b0cb037ba79bc489ee1fcf211d72c4af3c433201d4c6599fddff11"} Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.123455 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-787bf4d777-x8gln" event={"ID":"06da071b-e1a1-42da-9e6e-fd957923876d","Type":"ContainerStarted","Data":"5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7"} Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.123493 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-787bf4d777-x8gln" event={"ID":"06da071b-e1a1-42da-9e6e-fd957923876d","Type":"ContainerStarted","Data":"8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70"} Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.123682 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.123834 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.150533 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-787bf4d777-x8gln" podStartSLOduration=9.150515121 podStartE2EDuration="9.150515121s" podCreationTimestamp="2025-10-04 03:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:50.146409663 +0000 UTC m=+1181.438419375" watchObservedRunningTime="2025-10-04 03:22:50.150515121 +0000 UTC m=+1181.442524833" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.218271 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.605997 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-pv8zj" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.765375 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrhrf\" (UniqueName: \"kubernetes.io/projected/cbd7579a-ef76-48fc-96f6-3b26466084f8-kube-api-access-hrhrf\") pod \"cbd7579a-ef76-48fc-96f6-3b26466084f8\" (UID: \"cbd7579a-ef76-48fc-96f6-3b26466084f8\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.775970 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbd7579a-ef76-48fc-96f6-3b26466084f8-kube-api-access-hrhrf" (OuterVolumeSpecName: "kube-api-access-hrhrf") pod "cbd7579a-ef76-48fc-96f6-3b26466084f8" (UID: "cbd7579a-ef76-48fc-96f6-3b26466084f8"). InnerVolumeSpecName "kube-api-access-hrhrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.793700 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kv9tv" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.822459 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.822894 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-n8mcv" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.869760 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px789\" (UniqueName: \"kubernetes.io/projected/23bf2c46-426f-4bb7-82e0-f9e90d0a2f32-kube-api-access-px789\") pod \"23bf2c46-426f-4bb7-82e0-f9e90d0a2f32\" (UID: \"23bf2c46-426f-4bb7-82e0-f9e90d0a2f32\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.870283 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrhrf\" (UniqueName: \"kubernetes.io/projected/cbd7579a-ef76-48fc-96f6-3b26466084f8-kube-api-access-hrhrf\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.876287 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23bf2c46-426f-4bb7-82e0-f9e90d0a2f32-kube-api-access-px789" (OuterVolumeSpecName: "kube-api-access-px789") pod "23bf2c46-426f-4bb7-82e0-f9e90d0a2f32" (UID: "23bf2c46-426f-4bb7-82e0-f9e90d0a2f32"). InnerVolumeSpecName "kube-api-access-px789". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.971797 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-config-data\") pod \"94883722-c05e-43f6-ac7d-91ce20de293b\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.971837 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-combined-ca-bundle\") pod \"94883722-c05e-43f6-ac7d-91ce20de293b\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.971893 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czxjf\" (UniqueName: \"kubernetes.io/projected/3c9fa8cd-207f-4561-9116-ceea539098d8-kube-api-access-czxjf\") pod \"3c9fa8cd-207f-4561-9116-ceea539098d8\" (UID: \"3c9fa8cd-207f-4561-9116-ceea539098d8\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.971913 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"94883722-c05e-43f6-ac7d-91ce20de293b\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.971934 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-logs\") pod \"94883722-c05e-43f6-ac7d-91ce20de293b\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.972000 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24vzb\" (UniqueName: \"kubernetes.io/projected/94883722-c05e-43f6-ac7d-91ce20de293b-kube-api-access-24vzb\") pod \"94883722-c05e-43f6-ac7d-91ce20de293b\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.972133 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-internal-tls-certs\") pod \"94883722-c05e-43f6-ac7d-91ce20de293b\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.972159 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-scripts\") pod \"94883722-c05e-43f6-ac7d-91ce20de293b\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.972191 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-httpd-run\") pod \"94883722-c05e-43f6-ac7d-91ce20de293b\" (UID: \"94883722-c05e-43f6-ac7d-91ce20de293b\") " Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.972523 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px789\" (UniqueName: \"kubernetes.io/projected/23bf2c46-426f-4bb7-82e0-f9e90d0a2f32-kube-api-access-px789\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.973497 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "94883722-c05e-43f6-ac7d-91ce20de293b" (UID: "94883722-c05e-43f6-ac7d-91ce20de293b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.973937 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-logs" (OuterVolumeSpecName: "logs") pod "94883722-c05e-43f6-ac7d-91ce20de293b" (UID: "94883722-c05e-43f6-ac7d-91ce20de293b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.979446 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94883722-c05e-43f6-ac7d-91ce20de293b-kube-api-access-24vzb" (OuterVolumeSpecName: "kube-api-access-24vzb") pod "94883722-c05e-43f6-ac7d-91ce20de293b" (UID: "94883722-c05e-43f6-ac7d-91ce20de293b"). InnerVolumeSpecName "kube-api-access-24vzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:50 crc kubenswrapper[4770]: I1004 03:22:50.979466 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "94883722-c05e-43f6-ac7d-91ce20de293b" (UID: "94883722-c05e-43f6-ac7d-91ce20de293b"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.013865 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c9fa8cd-207f-4561-9116-ceea539098d8-kube-api-access-czxjf" (OuterVolumeSpecName: "kube-api-access-czxjf") pod "3c9fa8cd-207f-4561-9116-ceea539098d8" (UID: "3c9fa8cd-207f-4561-9116-ceea539098d8"). InnerVolumeSpecName "kube-api-access-czxjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.014756 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-scripts" (OuterVolumeSpecName: "scripts") pod "94883722-c05e-43f6-ac7d-91ce20de293b" (UID: "94883722-c05e-43f6-ac7d-91ce20de293b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.039139 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94883722-c05e-43f6-ac7d-91ce20de293b" (UID: "94883722-c05e-43f6-ac7d-91ce20de293b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.048408 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "94883722-c05e-43f6-ac7d-91ce20de293b" (UID: "94883722-c05e-43f6-ac7d-91ce20de293b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.061772 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-config-data" (OuterVolumeSpecName: "config-data") pod "94883722-c05e-43f6-ac7d-91ce20de293b" (UID: "94883722-c05e-43f6-ac7d-91ce20de293b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.073934 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.073977 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.073986 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.073994 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.074024 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94883722-c05e-43f6-ac7d-91ce20de293b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.074058 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.074068 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czxjf\" (UniqueName: \"kubernetes.io/projected/3c9fa8cd-207f-4561-9116-ceea539098d8-kube-api-access-czxjf\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.074080 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94883722-c05e-43f6-ac7d-91ce20de293b-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.074109 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24vzb\" (UniqueName: \"kubernetes.io/projected/94883722-c05e-43f6-ac7d-91ce20de293b-kube-api-access-24vzb\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.096635 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.136922 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d","Type":"ContainerStarted","Data":"b945952abce048f596643629d0b14f695125f7fc6eddcf6aec9901c2860f9cbb"} Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.139277 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-pv8zj" event={"ID":"cbd7579a-ef76-48fc-96f6-3b26466084f8","Type":"ContainerDied","Data":"687f46dfbfb423c36674fba2ad86d4252644374bb8a33659bd602108047a4520"} Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.139296 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="687f46dfbfb423c36674fba2ad86d4252644374bb8a33659bd602108047a4520" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.139356 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-pv8zj" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.142831 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kv9tv" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.142875 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kv9tv" event={"ID":"23bf2c46-426f-4bb7-82e0-f9e90d0a2f32","Type":"ContainerDied","Data":"59d4338ba8f0336c488c1c0fd551f96cb38777c8f9ffde34fb2b9cf9a7b1889c"} Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.142931 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59d4338ba8f0336c488c1c0fd551f96cb38777c8f9ffde34fb2b9cf9a7b1889c" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.144565 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-n8mcv" event={"ID":"3c9fa8cd-207f-4561-9116-ceea539098d8","Type":"ContainerDied","Data":"58bfdd93d512adbe87b4be0e04e74adfee679c60d791c8c3e70c035d7608b780"} Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.144600 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58bfdd93d512adbe87b4be0e04e74adfee679c60d791c8c3e70c035d7608b780" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.144669 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-n8mcv" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.161343 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94883722-c05e-43f6-ac7d-91ce20de293b","Type":"ContainerDied","Data":"469ac5cef90b33bfd29d462148a87e94a98f7d1e15639d4dec3baf0f3aacbda7"} Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.161413 4770 scope.go:117] "RemoveContainer" containerID="9d13162cc33717d452f99db8ca695a367a80b9d5bc0d24acc6684388fef67953" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.161781 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.173521 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerStarted","Data":"9f61b18862458bd5a59b592c737d6785024ea2b935d2c97771e3c1dfde2196f7"} Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.175830 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.217092 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.223677 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257078 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:22:51 crc kubenswrapper[4770]: E1004 03:22:51.257581 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94883722-c05e-43f6-ac7d-91ce20de293b" containerName="glance-log" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257606 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="94883722-c05e-43f6-ac7d-91ce20de293b" containerName="glance-log" Oct 04 03:22:51 crc kubenswrapper[4770]: E1004 03:22:51.257634 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c9fa8cd-207f-4561-9116-ceea539098d8" containerName="mariadb-database-create" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257644 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c9fa8cd-207f-4561-9116-ceea539098d8" containerName="mariadb-database-create" Oct 04 03:22:51 crc kubenswrapper[4770]: E1004 03:22:51.257655 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbd7579a-ef76-48fc-96f6-3b26466084f8" containerName="mariadb-database-create" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257663 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbd7579a-ef76-48fc-96f6-3b26466084f8" containerName="mariadb-database-create" Oct 04 03:22:51 crc kubenswrapper[4770]: E1004 03:22:51.257678 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94883722-c05e-43f6-ac7d-91ce20de293b" containerName="glance-httpd" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257687 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="94883722-c05e-43f6-ac7d-91ce20de293b" containerName="glance-httpd" Oct 04 03:22:51 crc kubenswrapper[4770]: E1004 03:22:51.257698 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23bf2c46-426f-4bb7-82e0-f9e90d0a2f32" containerName="mariadb-database-create" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257705 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="23bf2c46-426f-4bb7-82e0-f9e90d0a2f32" containerName="mariadb-database-create" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257900 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="94883722-c05e-43f6-ac7d-91ce20de293b" containerName="glance-log" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257913 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbd7579a-ef76-48fc-96f6-3b26466084f8" containerName="mariadb-database-create" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257930 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c9fa8cd-207f-4561-9116-ceea539098d8" containerName="mariadb-database-create" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257950 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="23bf2c46-426f-4bb7-82e0-f9e90d0a2f32" containerName="mariadb-database-create" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.257967 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="94883722-c05e-43f6-ac7d-91ce20de293b" containerName="glance-httpd" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.259098 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.263106 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.263904 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.272658 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.293143 4770 scope.go:117] "RemoveContainer" containerID="36527a07ae6792b10bdaf728b57e11696b53baf86da649a634ef2e860974b76b" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.379603 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.379877 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-logs\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.379991 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.380129 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.380255 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.380348 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lmb9\" (UniqueName: \"kubernetes.io/projected/d49edb80-7b9b-4b10-afbc-652ec540542e-kube-api-access-5lmb9\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.380422 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.380546 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.482678 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.482740 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.482783 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-logs\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.482808 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.482830 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.482898 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.482927 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lmb9\" (UniqueName: \"kubernetes.io/projected/d49edb80-7b9b-4b10-afbc-652ec540542e-kube-api-access-5lmb9\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.482946 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.483923 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.483949 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.484254 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-logs\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.488787 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.490547 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.496186 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.498556 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.517737 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lmb9\" (UniqueName: \"kubernetes.io/projected/d49edb80-7b9b-4b10-afbc-652ec540542e-kube-api-access-5lmb9\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.574413 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.592781 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:22:51 crc kubenswrapper[4770]: I1004 03:22:51.726958 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94883722-c05e-43f6-ac7d-91ce20de293b" path="/var/lib/kubelet/pods/94883722-c05e-43f6-ac7d-91ce20de293b/volumes" Oct 04 03:22:52 crc kubenswrapper[4770]: I1004 03:22:52.324017 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:22:52 crc kubenswrapper[4770]: W1004 03:22:52.327178 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd49edb80_7b9b_4b10_afbc_652ec540542e.slice/crio-d6066afec7af3ac77155b530eef0a09aaada0e0f6fb17bf3d05343aaad18f073 WatchSource:0}: Error finding container d6066afec7af3ac77155b530eef0a09aaada0e0f6fb17bf3d05343aaad18f073: Status 404 returned error can't find the container with id d6066afec7af3ac77155b530eef0a09aaada0e0f6fb17bf3d05343aaad18f073 Oct 04 03:22:52 crc kubenswrapper[4770]: I1004 03:22:52.479727 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.210713 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d","Type":"ContainerStarted","Data":"1f8470160c79b09e183b8b0818cd1ff1775ade9ca40792f2aa78671ca11f3a5b"} Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.216188 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d49edb80-7b9b-4b10-afbc-652ec540542e","Type":"ContainerStarted","Data":"e8c9ef6e4c5a603a6c8991200968c6b52e850056eec488d0768757430caac141"} Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.216436 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d49edb80-7b9b-4b10-afbc-652ec540542e","Type":"ContainerStarted","Data":"d6066afec7af3ac77155b530eef0a09aaada0e0f6fb17bf3d05343aaad18f073"} Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.219440 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerStarted","Data":"c5afe1c08a4324a4d3d38d864ce48d876264c35e92f6b5f2cad3f25fce355b35"} Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.235490 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.23546877 podStartE2EDuration="5.23546877s" podCreationTimestamp="2025-10-04 03:22:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:53.234064873 +0000 UTC m=+1184.526074575" watchObservedRunningTime="2025-10-04 03:22:53.23546877 +0000 UTC m=+1184.527478482" Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.380815 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3321-account-create-zxjrn"] Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.383738 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3321-account-create-zxjrn" Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.387157 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.413273 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3321-account-create-zxjrn"] Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.441544 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb2x7\" (UniqueName: \"kubernetes.io/projected/45047928-b458-48d5-b62a-ab2bd91d6adb-kube-api-access-xb2x7\") pod \"nova-api-3321-account-create-zxjrn\" (UID: \"45047928-b458-48d5-b62a-ab2bd91d6adb\") " pod="openstack/nova-api-3321-account-create-zxjrn" Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.543625 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb2x7\" (UniqueName: \"kubernetes.io/projected/45047928-b458-48d5-b62a-ab2bd91d6adb-kube-api-access-xb2x7\") pod \"nova-api-3321-account-create-zxjrn\" (UID: \"45047928-b458-48d5-b62a-ab2bd91d6adb\") " pod="openstack/nova-api-3321-account-create-zxjrn" Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.580533 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb2x7\" (UniqueName: \"kubernetes.io/projected/45047928-b458-48d5-b62a-ab2bd91d6adb-kube-api-access-xb2x7\") pod \"nova-api-3321-account-create-zxjrn\" (UID: \"45047928-b458-48d5-b62a-ab2bd91d6adb\") " pod="openstack/nova-api-3321-account-create-zxjrn" Oct 04 03:22:53 crc kubenswrapper[4770]: I1004 03:22:53.724694 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3321-account-create-zxjrn" Oct 04 03:22:54 crc kubenswrapper[4770]: I1004 03:22:54.231078 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerStarted","Data":"d2802c6f1e56f98dff3b354d0c688c189b8cffec9a029495e21ae89dc45888c5"} Oct 04 03:22:54 crc kubenswrapper[4770]: I1004 03:22:54.233200 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d49edb80-7b9b-4b10-afbc-652ec540542e","Type":"ContainerStarted","Data":"4e30affad8c19685e28c648ecf8ff6c026d5dda9129c4944f788d2e762582b04"} Oct 04 03:22:54 crc kubenswrapper[4770]: I1004 03:22:54.257814 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.257792846 podStartE2EDuration="3.257792846s" podCreationTimestamp="2025-10-04 03:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:22:54.253575436 +0000 UTC m=+1185.545585158" watchObservedRunningTime="2025-10-04 03:22:54.257792846 +0000 UTC m=+1185.549802558" Oct 04 03:22:54 crc kubenswrapper[4770]: I1004 03:22:54.313082 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3321-account-create-zxjrn"] Oct 04 03:22:54 crc kubenswrapper[4770]: W1004 03:22:54.322336 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45047928_b458_48d5_b62a_ab2bd91d6adb.slice/crio-d09024b4353af444202906b37f9c90e5fd11bd3fd11ddfeef4796e46794315bd WatchSource:0}: Error finding container d09024b4353af444202906b37f9c90e5fd11bd3fd11ddfeef4796e46794315bd: Status 404 returned error can't find the container with id d09024b4353af444202906b37f9c90e5fd11bd3fd11ddfeef4796e46794315bd Oct 04 03:22:55 crc kubenswrapper[4770]: I1004 03:22:55.251625 4770 generic.go:334] "Generic (PLEG): container finished" podID="45047928-b458-48d5-b62a-ab2bd91d6adb" containerID="26f529f3c164e5da28d60a276c95d6939aee3b026f95522050e7f3898614759a" exitCode=0 Oct 04 03:22:55 crc kubenswrapper[4770]: I1004 03:22:55.251746 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3321-account-create-zxjrn" event={"ID":"45047928-b458-48d5-b62a-ab2bd91d6adb","Type":"ContainerDied","Data":"26f529f3c164e5da28d60a276c95d6939aee3b026f95522050e7f3898614759a"} Oct 04 03:22:55 crc kubenswrapper[4770]: I1004 03:22:55.252055 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3321-account-create-zxjrn" event={"ID":"45047928-b458-48d5-b62a-ab2bd91d6adb","Type":"ContainerStarted","Data":"d09024b4353af444202906b37f9c90e5fd11bd3fd11ddfeef4796e46794315bd"} Oct 04 03:22:55 crc kubenswrapper[4770]: I1004 03:22:55.255411 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerStarted","Data":"67d5d43b0f027c265051875c649556b3651ea5ed25eb21d93f27424f9cb3af2f"} Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.277583 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerStarted","Data":"199a4344a50d40a792317c1d70a34175f51ddc57100e429b9e71018922ce3a8b"} Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.277792 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="ceilometer-central-agent" containerID="cri-o://c5afe1c08a4324a4d3d38d864ce48d876264c35e92f6b5f2cad3f25fce355b35" gracePeriod=30 Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.277863 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="proxy-httpd" containerID="cri-o://199a4344a50d40a792317c1d70a34175f51ddc57100e429b9e71018922ce3a8b" gracePeriod=30 Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.277907 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="ceilometer-notification-agent" containerID="cri-o://d2802c6f1e56f98dff3b354d0c688c189b8cffec9a029495e21ae89dc45888c5" gracePeriod=30 Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.278002 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="sg-core" containerID="cri-o://67d5d43b0f027c265051875c649556b3651ea5ed25eb21d93f27424f9cb3af2f" gracePeriod=30 Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.331492 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.028437429 podStartE2EDuration="7.331427387s" podCreationTimestamp="2025-10-04 03:22:49 +0000 UTC" firstStartedPulling="2025-10-04 03:22:50.2701207 +0000 UTC m=+1181.562130422" lastFinishedPulling="2025-10-04 03:22:55.573110678 +0000 UTC m=+1186.865120380" observedRunningTime="2025-10-04 03:22:56.315515901 +0000 UTC m=+1187.607525613" watchObservedRunningTime="2025-10-04 03:22:56.331427387 +0000 UTC m=+1187.623437109" Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.667924 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.686918 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.770245 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3321-account-create-zxjrn" Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.833711 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb2x7\" (UniqueName: \"kubernetes.io/projected/45047928-b458-48d5-b62a-ab2bd91d6adb-kube-api-access-xb2x7\") pod \"45047928-b458-48d5-b62a-ab2bd91d6adb\" (UID: \"45047928-b458-48d5-b62a-ab2bd91d6adb\") " Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.841393 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45047928-b458-48d5-b62a-ab2bd91d6adb-kube-api-access-xb2x7" (OuterVolumeSpecName: "kube-api-access-xb2x7") pod "45047928-b458-48d5-b62a-ab2bd91d6adb" (UID: "45047928-b458-48d5-b62a-ab2bd91d6adb"). InnerVolumeSpecName "kube-api-access-xb2x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:56 crc kubenswrapper[4770]: I1004 03:22:56.937609 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb2x7\" (UniqueName: \"kubernetes.io/projected/45047928-b458-48d5-b62a-ab2bd91d6adb-kube-api-access-xb2x7\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.291165 4770 generic.go:334] "Generic (PLEG): container finished" podID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerID="199a4344a50d40a792317c1d70a34175f51ddc57100e429b9e71018922ce3a8b" exitCode=0 Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.291242 4770 generic.go:334] "Generic (PLEG): container finished" podID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerID="67d5d43b0f027c265051875c649556b3651ea5ed25eb21d93f27424f9cb3af2f" exitCode=2 Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.291255 4770 generic.go:334] "Generic (PLEG): container finished" podID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerID="d2802c6f1e56f98dff3b354d0c688c189b8cffec9a029495e21ae89dc45888c5" exitCode=0 Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.291266 4770 generic.go:334] "Generic (PLEG): container finished" podID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerID="c5afe1c08a4324a4d3d38d864ce48d876264c35e92f6b5f2cad3f25fce355b35" exitCode=0 Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.291336 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerDied","Data":"199a4344a50d40a792317c1d70a34175f51ddc57100e429b9e71018922ce3a8b"} Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.291400 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerDied","Data":"67d5d43b0f027c265051875c649556b3651ea5ed25eb21d93f27424f9cb3af2f"} Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.291417 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerDied","Data":"d2802c6f1e56f98dff3b354d0c688c189b8cffec9a029495e21ae89dc45888c5"} Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.291430 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerDied","Data":"c5afe1c08a4324a4d3d38d864ce48d876264c35e92f6b5f2cad3f25fce355b35"} Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.293716 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3321-account-create-zxjrn" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.294396 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3321-account-create-zxjrn" event={"ID":"45047928-b458-48d5-b62a-ab2bd91d6adb","Type":"ContainerDied","Data":"d09024b4353af444202906b37f9c90e5fd11bd3fd11ddfeef4796e46794315bd"} Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.294422 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d09024b4353af444202906b37f9c90e5fd11bd3fd11ddfeef4796e46794315bd" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.334501 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.470556 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-scripts\") pod \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.470991 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-combined-ca-bundle\") pod \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.471110 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9c2z\" (UniqueName: \"kubernetes.io/projected/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-kube-api-access-j9c2z\") pod \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.471164 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-sg-core-conf-yaml\") pod \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.471300 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-config-data\") pod \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.471328 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-log-httpd\") pod \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.471371 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-run-httpd\") pod \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\" (UID: \"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53\") " Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.471986 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" (UID: "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.472133 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" (UID: "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.482257 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-scripts" (OuterVolumeSpecName: "scripts") pod "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" (UID: "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.482347 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-kube-api-access-j9c2z" (OuterVolumeSpecName: "kube-api-access-j9c2z") pod "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" (UID: "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53"). InnerVolumeSpecName "kube-api-access-j9c2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.505951 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" (UID: "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.568621 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" (UID: "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.574600 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9c2z\" (UniqueName: \"kubernetes.io/projected/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-kube-api-access-j9c2z\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.574635 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.574649 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.574661 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.574670 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.574724 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.585255 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-config-data" (OuterVolumeSpecName: "config-data") pod "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" (UID: "ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:22:57 crc kubenswrapper[4770]: I1004 03:22:57.676056 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.323532 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53","Type":"ContainerDied","Data":"9f61b18862458bd5a59b592c737d6785024ea2b935d2c97771e3c1dfde2196f7"} Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.323625 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.323632 4770 scope.go:117] "RemoveContainer" containerID="199a4344a50d40a792317c1d70a34175f51ddc57100e429b9e71018922ce3a8b" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.353583 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.366453 4770 scope.go:117] "RemoveContainer" containerID="67d5d43b0f027c265051875c649556b3651ea5ed25eb21d93f27424f9cb3af2f" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.381563 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.393464 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:58 crc kubenswrapper[4770]: E1004 03:22:58.393906 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="ceilometer-central-agent" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.393920 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="ceilometer-central-agent" Oct 04 03:22:58 crc kubenswrapper[4770]: E1004 03:22:58.393929 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="sg-core" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.393937 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="sg-core" Oct 04 03:22:58 crc kubenswrapper[4770]: E1004 03:22:58.393947 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="ceilometer-notification-agent" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.393953 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="ceilometer-notification-agent" Oct 04 03:22:58 crc kubenswrapper[4770]: E1004 03:22:58.393964 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45047928-b458-48d5-b62a-ab2bd91d6adb" containerName="mariadb-account-create" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.393970 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="45047928-b458-48d5-b62a-ab2bd91d6adb" containerName="mariadb-account-create" Oct 04 03:22:58 crc kubenswrapper[4770]: E1004 03:22:58.393986 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="proxy-httpd" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.393991 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="proxy-httpd" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.394194 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="45047928-b458-48d5-b62a-ab2bd91d6adb" containerName="mariadb-account-create" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.394207 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="ceilometer-central-agent" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.394218 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="proxy-httpd" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.394231 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="sg-core" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.394245 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" containerName="ceilometer-notification-agent" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.395896 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.405914 4770 scope.go:117] "RemoveContainer" containerID="d2802c6f1e56f98dff3b354d0c688c189b8cffec9a029495e21ae89dc45888c5" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.406756 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.407278 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.410273 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.437141 4770 scope.go:117] "RemoveContainer" containerID="c5afe1c08a4324a4d3d38d864ce48d876264c35e92f6b5f2cad3f25fce355b35" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.489164 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.489214 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-log-httpd\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.489422 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsv4m\" (UniqueName: \"kubernetes.io/projected/be265de0-637c-43e8-ab8d-60eff6e1d55b-kube-api-access-wsv4m\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.489511 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-config-data\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.489592 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.489886 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-scripts\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.492931 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-run-httpd\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.595825 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-config-data\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.595911 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.595956 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-scripts\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.595985 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-run-httpd\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.596070 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.596103 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-log-httpd\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.596231 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsv4m\" (UniqueName: \"kubernetes.io/projected/be265de0-637c-43e8-ab8d-60eff6e1d55b-kube-api-access-wsv4m\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.596571 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-run-httpd\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.596954 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-log-httpd\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.603891 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.604427 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-scripts\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.613772 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.628830 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-config-data\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.648569 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsv4m\" (UniqueName: \"kubernetes.io/projected/be265de0-637c-43e8-ab8d-60eff6e1d55b-kube-api-access-wsv4m\") pod \"ceilometer-0\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.724390 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.724438 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.729326 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.768249 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 03:22:58 crc kubenswrapper[4770]: I1004 03:22:58.772691 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 03:22:59 crc kubenswrapper[4770]: I1004 03:22:59.211214 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:22:59 crc kubenswrapper[4770]: I1004 03:22:59.333799 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerStarted","Data":"57263622b06bbcd6c5525e8f308042ebde0cd3c3f3fbeced3d888e731c531b14"} Oct 04 03:22:59 crc kubenswrapper[4770]: I1004 03:22:59.336331 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 03:22:59 crc kubenswrapper[4770]: I1004 03:22:59.336376 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 03:22:59 crc kubenswrapper[4770]: I1004 03:22:59.689445 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53" path="/var/lib/kubelet/pods/ccbd0d9e-7a8a-47cc-bdd1-4cd84c4e2b53/volumes" Oct 04 03:23:00 crc kubenswrapper[4770]: I1004 03:23:00.356557 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerStarted","Data":"647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2"} Oct 04 03:23:00 crc kubenswrapper[4770]: I1004 03:23:00.417701 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:23:01 crc kubenswrapper[4770]: I1004 03:23:01.291553 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 03:23:01 crc kubenswrapper[4770]: I1004 03:23:01.365235 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 03:23:01 crc kubenswrapper[4770]: I1004 03:23:01.368288 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerStarted","Data":"325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20"} Oct 04 03:23:01 crc kubenswrapper[4770]: I1004 03:23:01.368323 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerStarted","Data":"9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a"} Oct 04 03:23:01 crc kubenswrapper[4770]: I1004 03:23:01.619469 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 03:23:01 crc kubenswrapper[4770]: I1004 03:23:01.620067 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 03:23:01 crc kubenswrapper[4770]: I1004 03:23:01.672400 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 03:23:01 crc kubenswrapper[4770]: I1004 03:23:01.695269 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 03:23:02 crc kubenswrapper[4770]: I1004 03:23:02.378785 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 03:23:02 crc kubenswrapper[4770]: I1004 03:23:02.380667 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.390667 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerStarted","Data":"ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf"} Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.392061 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.422643 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.107799886 podStartE2EDuration="5.42262132s" podCreationTimestamp="2025-10-04 03:22:58 +0000 UTC" firstStartedPulling="2025-10-04 03:22:59.234666623 +0000 UTC m=+1190.526676355" lastFinishedPulling="2025-10-04 03:23:02.549488077 +0000 UTC m=+1193.841497789" observedRunningTime="2025-10-04 03:23:03.417891136 +0000 UTC m=+1194.709900858" watchObservedRunningTime="2025-10-04 03:23:03.42262132 +0000 UTC m=+1194.714631032" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.517058 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-a2a5-account-create-d6vtx"] Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.519709 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a2a5-account-create-d6vtx" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.524818 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.558844 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a2a5-account-create-d6vtx"] Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.606543 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.697794 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5dbdffbfb6-pnbvk"] Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.698021 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5dbdffbfb6-pnbvk" podUID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerName="neutron-api" containerID="cri-o://456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75" gracePeriod=30 Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.698320 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5dbdffbfb6-pnbvk" podUID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerName="neutron-httpd" containerID="cri-o://e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d" gracePeriod=30 Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.708227 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqxgj\" (UniqueName: \"kubernetes.io/projected/a1deee0b-c346-457f-94e1-7480502d0583-kube-api-access-nqxgj\") pod \"nova-cell0-a2a5-account-create-d6vtx\" (UID: \"a1deee0b-c346-457f-94e1-7480502d0583\") " pod="openstack/nova-cell0-a2a5-account-create-d6vtx" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.714110 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-09d5-account-create-5kncn"] Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.715275 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-09d5-account-create-5kncn" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.719546 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.725787 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-09d5-account-create-5kncn"] Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.810119 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqxgj\" (UniqueName: \"kubernetes.io/projected/a1deee0b-c346-457f-94e1-7480502d0583-kube-api-access-nqxgj\") pod \"nova-cell0-a2a5-account-create-d6vtx\" (UID: \"a1deee0b-c346-457f-94e1-7480502d0583\") " pod="openstack/nova-cell0-a2a5-account-create-d6vtx" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.837851 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqxgj\" (UniqueName: \"kubernetes.io/projected/a1deee0b-c346-457f-94e1-7480502d0583-kube-api-access-nqxgj\") pod \"nova-cell0-a2a5-account-create-d6vtx\" (UID: \"a1deee0b-c346-457f-94e1-7480502d0583\") " pod="openstack/nova-cell0-a2a5-account-create-d6vtx" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.853584 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a2a5-account-create-d6vtx" Oct 04 03:23:03 crc kubenswrapper[4770]: I1004 03:23:03.911740 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwfjq\" (UniqueName: \"kubernetes.io/projected/b417baa3-f6c2-499f-af94-0bd91ff07e61-kube-api-access-kwfjq\") pod \"nova-cell1-09d5-account-create-5kncn\" (UID: \"b417baa3-f6c2-499f-af94-0bd91ff07e61\") " pod="openstack/nova-cell1-09d5-account-create-5kncn" Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.015585 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwfjq\" (UniqueName: \"kubernetes.io/projected/b417baa3-f6c2-499f-af94-0bd91ff07e61-kube-api-access-kwfjq\") pod \"nova-cell1-09d5-account-create-5kncn\" (UID: \"b417baa3-f6c2-499f-af94-0bd91ff07e61\") " pod="openstack/nova-cell1-09d5-account-create-5kncn" Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.044780 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwfjq\" (UniqueName: \"kubernetes.io/projected/b417baa3-f6c2-499f-af94-0bd91ff07e61-kube-api-access-kwfjq\") pod \"nova-cell1-09d5-account-create-5kncn\" (UID: \"b417baa3-f6c2-499f-af94-0bd91ff07e61\") " pod="openstack/nova-cell1-09d5-account-create-5kncn" Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.046740 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-09d5-account-create-5kncn" Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.404742 4770 generic.go:334] "Generic (PLEG): container finished" podID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerID="e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d" exitCode=0 Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.405217 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dbdffbfb6-pnbvk" event={"ID":"43a28d77-6122-4774-b40d-f92dfdc9933d","Type":"ContainerDied","Data":"e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d"} Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.405319 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.405328 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.489600 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a2a5-account-create-d6vtx"] Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.592514 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-09d5-account-create-5kncn"] Oct 04 03:23:04 crc kubenswrapper[4770]: I1004 03:23:04.858286 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.422230 4770 generic.go:334] "Generic (PLEG): container finished" podID="a1deee0b-c346-457f-94e1-7480502d0583" containerID="158bcc32260df183555b1a09c48ad264d70fd241b7a7bab2d1ed22f5fd911d0a" exitCode=0 Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.422643 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a2a5-account-create-d6vtx" event={"ID":"a1deee0b-c346-457f-94e1-7480502d0583","Type":"ContainerDied","Data":"158bcc32260df183555b1a09c48ad264d70fd241b7a7bab2d1ed22f5fd911d0a"} Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.422676 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a2a5-account-create-d6vtx" event={"ID":"a1deee0b-c346-457f-94e1-7480502d0583","Type":"ContainerStarted","Data":"84c0a82117835fbbcf019d0f1cc1598573c5104ea220046191dae2d58eca93c0"} Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.440241 4770 generic.go:334] "Generic (PLEG): container finished" podID="b417baa3-f6c2-499f-af94-0bd91ff07e61" containerID="7faf1c073ac70ed8106d0926d77195ec92cf251f3043e92547e6382b188686fb" exitCode=0 Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.440337 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.441233 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-09d5-account-create-5kncn" event={"ID":"b417baa3-f6c2-499f-af94-0bd91ff07e61","Type":"ContainerDied","Data":"7faf1c073ac70ed8106d0926d77195ec92cf251f3043e92547e6382b188686fb"} Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.441261 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-09d5-account-create-5kncn" event={"ID":"b417baa3-f6c2-499f-af94-0bd91ff07e61","Type":"ContainerStarted","Data":"d003823873eb13a9f77691e9f6d78322c81705f8bf44b9b74c89bf37354cfb66"} Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.449763 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.450044 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="ceilometer-central-agent" containerID="cri-o://647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2" gracePeriod=30 Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.450173 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="ceilometer-notification-agent" containerID="cri-o://9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a" gracePeriod=30 Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.450212 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="proxy-httpd" containerID="cri-o://ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf" gracePeriod=30 Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.450181 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="sg-core" containerID="cri-o://325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20" gracePeriod=30 Oct 04 03:23:05 crc kubenswrapper[4770]: I1004 03:23:05.567382 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 03:23:06 crc kubenswrapper[4770]: I1004 03:23:06.451576 4770 generic.go:334] "Generic (PLEG): container finished" podID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerID="ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf" exitCode=0 Oct 04 03:23:06 crc kubenswrapper[4770]: I1004 03:23:06.451909 4770 generic.go:334] "Generic (PLEG): container finished" podID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerID="325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20" exitCode=2 Oct 04 03:23:06 crc kubenswrapper[4770]: I1004 03:23:06.451922 4770 generic.go:334] "Generic (PLEG): container finished" podID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerID="9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a" exitCode=0 Oct 04 03:23:06 crc kubenswrapper[4770]: I1004 03:23:06.451733 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerDied","Data":"ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf"} Oct 04 03:23:06 crc kubenswrapper[4770]: I1004 03:23:06.452121 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerDied","Data":"325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20"} Oct 04 03:23:06 crc kubenswrapper[4770]: I1004 03:23:06.452135 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerDied","Data":"9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a"} Oct 04 03:23:06 crc kubenswrapper[4770]: I1004 03:23:06.920408 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a2a5-account-create-d6vtx" Oct 04 03:23:06 crc kubenswrapper[4770]: I1004 03:23:06.925969 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-09d5-account-create-5kncn" Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.072951 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqxgj\" (UniqueName: \"kubernetes.io/projected/a1deee0b-c346-457f-94e1-7480502d0583-kube-api-access-nqxgj\") pod \"a1deee0b-c346-457f-94e1-7480502d0583\" (UID: \"a1deee0b-c346-457f-94e1-7480502d0583\") " Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.073046 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwfjq\" (UniqueName: \"kubernetes.io/projected/b417baa3-f6c2-499f-af94-0bd91ff07e61-kube-api-access-kwfjq\") pod \"b417baa3-f6c2-499f-af94-0bd91ff07e61\" (UID: \"b417baa3-f6c2-499f-af94-0bd91ff07e61\") " Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.081313 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1deee0b-c346-457f-94e1-7480502d0583-kube-api-access-nqxgj" (OuterVolumeSpecName: "kube-api-access-nqxgj") pod "a1deee0b-c346-457f-94e1-7480502d0583" (UID: "a1deee0b-c346-457f-94e1-7480502d0583"). InnerVolumeSpecName "kube-api-access-nqxgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.082567 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b417baa3-f6c2-499f-af94-0bd91ff07e61-kube-api-access-kwfjq" (OuterVolumeSpecName: "kube-api-access-kwfjq") pod "b417baa3-f6c2-499f-af94-0bd91ff07e61" (UID: "b417baa3-f6c2-499f-af94-0bd91ff07e61"). InnerVolumeSpecName "kube-api-access-kwfjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.175853 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqxgj\" (UniqueName: \"kubernetes.io/projected/a1deee0b-c346-457f-94e1-7480502d0583-kube-api-access-nqxgj\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.175891 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwfjq\" (UniqueName: \"kubernetes.io/projected/b417baa3-f6c2-499f-af94-0bd91ff07e61-kube-api-access-kwfjq\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.463676 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-09d5-account-create-5kncn" event={"ID":"b417baa3-f6c2-499f-af94-0bd91ff07e61","Type":"ContainerDied","Data":"d003823873eb13a9f77691e9f6d78322c81705f8bf44b9b74c89bf37354cfb66"} Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.463957 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d003823873eb13a9f77691e9f6d78322c81705f8bf44b9b74c89bf37354cfb66" Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.463703 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-09d5-account-create-5kncn" Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.466280 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a2a5-account-create-d6vtx" event={"ID":"a1deee0b-c346-457f-94e1-7480502d0583","Type":"ContainerDied","Data":"84c0a82117835fbbcf019d0f1cc1598573c5104ea220046191dae2d58eca93c0"} Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.466318 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a2a5-account-create-d6vtx" Oct 04 03:23:07 crc kubenswrapper[4770]: I1004 03:23:07.466323 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84c0a82117835fbbcf019d0f1cc1598573c5104ea220046191dae2d58eca93c0" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.438128 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.476893 4770 generic.go:334] "Generic (PLEG): container finished" podID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerID="456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75" exitCode=0 Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.476937 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dbdffbfb6-pnbvk" event={"ID":"43a28d77-6122-4774-b40d-f92dfdc9933d","Type":"ContainerDied","Data":"456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75"} Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.476966 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dbdffbfb6-pnbvk" event={"ID":"43a28d77-6122-4774-b40d-f92dfdc9933d","Type":"ContainerDied","Data":"bc0c63b76f969dcd85bf1b84c4e7bf1821bdbe5c062f0a685fe453e0e0b37185"} Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.476985 4770 scope.go:117] "RemoveContainer" containerID="e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.477131 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dbdffbfb6-pnbvk" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.508553 4770 scope.go:117] "RemoveContainer" containerID="456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.527319 4770 scope.go:117] "RemoveContainer" containerID="e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d" Oct 04 03:23:08 crc kubenswrapper[4770]: E1004 03:23:08.527802 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d\": container with ID starting with e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d not found: ID does not exist" containerID="e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.527835 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d"} err="failed to get container status \"e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d\": rpc error: code = NotFound desc = could not find container \"e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d\": container with ID starting with e0240e42ac4a0e61ee6d304194d0db7efde6e17e3e49d7a721b5b87065b2e58d not found: ID does not exist" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.527861 4770 scope.go:117] "RemoveContainer" containerID="456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75" Oct 04 03:23:08 crc kubenswrapper[4770]: E1004 03:23:08.528143 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75\": container with ID starting with 456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75 not found: ID does not exist" containerID="456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.528164 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75"} err="failed to get container status \"456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75\": rpc error: code = NotFound desc = could not find container \"456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75\": container with ID starting with 456818c68134469da6190fa81a309c306ea2a16d8bf2fc658f0802d3e5a42a75 not found: ID does not exist" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.602531 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-config\") pod \"43a28d77-6122-4774-b40d-f92dfdc9933d\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.602836 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-ovndb-tls-certs\") pod \"43a28d77-6122-4774-b40d-f92dfdc9933d\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.602864 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-httpd-config\") pod \"43a28d77-6122-4774-b40d-f92dfdc9933d\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.603431 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rzdb\" (UniqueName: \"kubernetes.io/projected/43a28d77-6122-4774-b40d-f92dfdc9933d-kube-api-access-9rzdb\") pod \"43a28d77-6122-4774-b40d-f92dfdc9933d\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.604745 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-combined-ca-bundle\") pod \"43a28d77-6122-4774-b40d-f92dfdc9933d\" (UID: \"43a28d77-6122-4774-b40d-f92dfdc9933d\") " Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.610996 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "43a28d77-6122-4774-b40d-f92dfdc9933d" (UID: "43a28d77-6122-4774-b40d-f92dfdc9933d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.613550 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43a28d77-6122-4774-b40d-f92dfdc9933d-kube-api-access-9rzdb" (OuterVolumeSpecName: "kube-api-access-9rzdb") pod "43a28d77-6122-4774-b40d-f92dfdc9933d" (UID: "43a28d77-6122-4774-b40d-f92dfdc9933d"). InnerVolumeSpecName "kube-api-access-9rzdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.667810 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43a28d77-6122-4774-b40d-f92dfdc9933d" (UID: "43a28d77-6122-4774-b40d-f92dfdc9933d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.672428 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-config" (OuterVolumeSpecName: "config") pod "43a28d77-6122-4774-b40d-f92dfdc9933d" (UID: "43a28d77-6122-4774-b40d-f92dfdc9933d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.710534 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.710581 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.710590 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.710601 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rzdb\" (UniqueName: \"kubernetes.io/projected/43a28d77-6122-4774-b40d-f92dfdc9933d-kube-api-access-9rzdb\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.730150 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "43a28d77-6122-4774-b40d-f92dfdc9933d" (UID: "43a28d77-6122-4774-b40d-f92dfdc9933d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.735680 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pz6zp"] Oct 04 03:23:08 crc kubenswrapper[4770]: E1004 03:23:08.736078 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b417baa3-f6c2-499f-af94-0bd91ff07e61" containerName="mariadb-account-create" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.736094 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b417baa3-f6c2-499f-af94-0bd91ff07e61" containerName="mariadb-account-create" Oct 04 03:23:08 crc kubenswrapper[4770]: E1004 03:23:08.736115 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1deee0b-c346-457f-94e1-7480502d0583" containerName="mariadb-account-create" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.736123 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1deee0b-c346-457f-94e1-7480502d0583" containerName="mariadb-account-create" Oct 04 03:23:08 crc kubenswrapper[4770]: E1004 03:23:08.736137 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerName="neutron-api" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.736143 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerName="neutron-api" Oct 04 03:23:08 crc kubenswrapper[4770]: E1004 03:23:08.736153 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerName="neutron-httpd" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.736158 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerName="neutron-httpd" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.736322 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerName="neutron-api" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.736349 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1deee0b-c346-457f-94e1-7480502d0583" containerName="mariadb-account-create" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.736360 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a28d77-6122-4774-b40d-f92dfdc9933d" containerName="neutron-httpd" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.736370 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b417baa3-f6c2-499f-af94-0bd91ff07e61" containerName="mariadb-account-create" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.736888 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.739114 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.739270 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.739409 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hdwhp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.745236 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pz6zp"] Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.814559 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.814651 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-scripts\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.814808 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-config-data\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.814895 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6n69\" (UniqueName: \"kubernetes.io/projected/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-kube-api-access-f6n69\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.815031 4770 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/43a28d77-6122-4774-b40d-f92dfdc9933d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.854067 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5dbdffbfb6-pnbvk"] Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.872866 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5dbdffbfb6-pnbvk"] Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.916197 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-config-data\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.916283 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6n69\" (UniqueName: \"kubernetes.io/projected/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-kube-api-access-f6n69\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.916386 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.916440 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-scripts\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.927781 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-config-data\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.935850 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.949514 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-scripts\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:08 crc kubenswrapper[4770]: I1004 03:23:08.955621 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6n69\" (UniqueName: \"kubernetes.io/projected/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-kube-api-access-f6n69\") pod \"nova-cell0-conductor-db-sync-pz6zp\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:09 crc kubenswrapper[4770]: I1004 03:23:09.095924 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:09 crc kubenswrapper[4770]: I1004 03:23:09.590842 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pz6zp"] Oct 04 03:23:09 crc kubenswrapper[4770]: W1004 03:23:09.594218 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacc1eb6c_ed84_4a06_a8da_c8c026f5c2a8.slice/crio-f9863f48804111d21fb32b80cc6a74de21f89da4257c2a405dcd9ca7072c39c5 WatchSource:0}: Error finding container f9863f48804111d21fb32b80cc6a74de21f89da4257c2a405dcd9ca7072c39c5: Status 404 returned error can't find the container with id f9863f48804111d21fb32b80cc6a74de21f89da4257c2a405dcd9ca7072c39c5 Oct 04 03:23:09 crc kubenswrapper[4770]: I1004 03:23:09.722866 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43a28d77-6122-4774-b40d-f92dfdc9933d" path="/var/lib/kubelet/pods/43a28d77-6122-4774-b40d-f92dfdc9933d/volumes" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.379121 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.444928 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-combined-ca-bundle\") pod \"be265de0-637c-43e8-ab8d-60eff6e1d55b\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.445147 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-sg-core-conf-yaml\") pod \"be265de0-637c-43e8-ab8d-60eff6e1d55b\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.445350 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-run-httpd\") pod \"be265de0-637c-43e8-ab8d-60eff6e1d55b\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.445415 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-config-data\") pod \"be265de0-637c-43e8-ab8d-60eff6e1d55b\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.445593 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsv4m\" (UniqueName: \"kubernetes.io/projected/be265de0-637c-43e8-ab8d-60eff6e1d55b-kube-api-access-wsv4m\") pod \"be265de0-637c-43e8-ab8d-60eff6e1d55b\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.445638 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-log-httpd\") pod \"be265de0-637c-43e8-ab8d-60eff6e1d55b\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.445684 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-scripts\") pod \"be265de0-637c-43e8-ab8d-60eff6e1d55b\" (UID: \"be265de0-637c-43e8-ab8d-60eff6e1d55b\") " Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.447654 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "be265de0-637c-43e8-ab8d-60eff6e1d55b" (UID: "be265de0-637c-43e8-ab8d-60eff6e1d55b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.450476 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "be265de0-637c-43e8-ab8d-60eff6e1d55b" (UID: "be265de0-637c-43e8-ab8d-60eff6e1d55b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.453622 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-scripts" (OuterVolumeSpecName: "scripts") pod "be265de0-637c-43e8-ab8d-60eff6e1d55b" (UID: "be265de0-637c-43e8-ab8d-60eff6e1d55b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.467384 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be265de0-637c-43e8-ab8d-60eff6e1d55b-kube-api-access-wsv4m" (OuterVolumeSpecName: "kube-api-access-wsv4m") pod "be265de0-637c-43e8-ab8d-60eff6e1d55b" (UID: "be265de0-637c-43e8-ab8d-60eff6e1d55b"). InnerVolumeSpecName "kube-api-access-wsv4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.491086 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "be265de0-637c-43e8-ab8d-60eff6e1d55b" (UID: "be265de0-637c-43e8-ab8d-60eff6e1d55b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.499163 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pz6zp" event={"ID":"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8","Type":"ContainerStarted","Data":"f9863f48804111d21fb32b80cc6a74de21f89da4257c2a405dcd9ca7072c39c5"} Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.503264 4770 generic.go:334] "Generic (PLEG): container finished" podID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerID="647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2" exitCode=0 Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.503309 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerDied","Data":"647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2"} Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.503339 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be265de0-637c-43e8-ab8d-60eff6e1d55b","Type":"ContainerDied","Data":"57263622b06bbcd6c5525e8f308042ebde0cd3c3f3fbeced3d888e731c531b14"} Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.503355 4770 scope.go:117] "RemoveContainer" containerID="ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.503578 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.535070 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be265de0-637c-43e8-ab8d-60eff6e1d55b" (UID: "be265de0-637c-43e8-ab8d-60eff6e1d55b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.548218 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsv4m\" (UniqueName: \"kubernetes.io/projected/be265de0-637c-43e8-ab8d-60eff6e1d55b-kube-api-access-wsv4m\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.548253 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.548266 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.548276 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.548286 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.548295 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be265de0-637c-43e8-ab8d-60eff6e1d55b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.568841 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-config-data" (OuterVolumeSpecName: "config-data") pod "be265de0-637c-43e8-ab8d-60eff6e1d55b" (UID: "be265de0-637c-43e8-ab8d-60eff6e1d55b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.652128 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be265de0-637c-43e8-ab8d-60eff6e1d55b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.672923 4770 scope.go:117] "RemoveContainer" containerID="325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.697294 4770 scope.go:117] "RemoveContainer" containerID="9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.726561 4770 scope.go:117] "RemoveContainer" containerID="647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.755481 4770 scope.go:117] "RemoveContainer" containerID="ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf" Oct 04 03:23:10 crc kubenswrapper[4770]: E1004 03:23:10.756055 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf\": container with ID starting with ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf not found: ID does not exist" containerID="ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.756097 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf"} err="failed to get container status \"ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf\": rpc error: code = NotFound desc = could not find container \"ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf\": container with ID starting with ca7863dc22ad04c1702d14efcfbf2bda9ebcb4bc7db89245e137300577d328bf not found: ID does not exist" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.756124 4770 scope.go:117] "RemoveContainer" containerID="325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20" Oct 04 03:23:10 crc kubenswrapper[4770]: E1004 03:23:10.756569 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20\": container with ID starting with 325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20 not found: ID does not exist" containerID="325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.756601 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20"} err="failed to get container status \"325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20\": rpc error: code = NotFound desc = could not find container \"325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20\": container with ID starting with 325648d198e24688be360f148351c40c490e475c1e74c8daa3f4f74c6231db20 not found: ID does not exist" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.756619 4770 scope.go:117] "RemoveContainer" containerID="9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a" Oct 04 03:23:10 crc kubenswrapper[4770]: E1004 03:23:10.756878 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a\": container with ID starting with 9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a not found: ID does not exist" containerID="9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.756907 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a"} err="failed to get container status \"9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a\": rpc error: code = NotFound desc = could not find container \"9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a\": container with ID starting with 9db4554dea3ed1cb4b262c7c3096a07ce41f5c04d6c90bc8e763635c9668c11a not found: ID does not exist" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.756926 4770 scope.go:117] "RemoveContainer" containerID="647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2" Oct 04 03:23:10 crc kubenswrapper[4770]: E1004 03:23:10.757273 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2\": container with ID starting with 647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2 not found: ID does not exist" containerID="647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.757301 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2"} err="failed to get container status \"647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2\": rpc error: code = NotFound desc = could not find container \"647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2\": container with ID starting with 647a9bb7492b1e58de59afb1712d1f85351a121115431a5b3af68e57d2af46b2 not found: ID does not exist" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.844636 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.865426 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.881777 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:10 crc kubenswrapper[4770]: E1004 03:23:10.882252 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="sg-core" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.882270 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="sg-core" Oct 04 03:23:10 crc kubenswrapper[4770]: E1004 03:23:10.882305 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="ceilometer-notification-agent" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.882312 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="ceilometer-notification-agent" Oct 04 03:23:10 crc kubenswrapper[4770]: E1004 03:23:10.882330 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="proxy-httpd" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.882338 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="proxy-httpd" Oct 04 03:23:10 crc kubenswrapper[4770]: E1004 03:23:10.882352 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="ceilometer-central-agent" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.882357 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="ceilometer-central-agent" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.882539 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="ceilometer-notification-agent" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.882568 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="proxy-httpd" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.882584 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="sg-core" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.882592 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" containerName="ceilometer-central-agent" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.884321 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.887432 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.887740 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.893257 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.957586 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.957642 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf5jt\" (UniqueName: \"kubernetes.io/projected/f2746b48-57ff-4848-9fd5-13862e6defd8-kube-api-access-nf5jt\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.958385 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-run-httpd\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.958457 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-config-data\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.958500 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-scripts\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.958720 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-log-httpd\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:10 crc kubenswrapper[4770]: I1004 03:23:10.959027 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.060707 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-run-httpd\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.060745 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-config-data\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.060774 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-scripts\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.060803 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-log-httpd\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.060892 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.060944 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.060970 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf5jt\" (UniqueName: \"kubernetes.io/projected/f2746b48-57ff-4848-9fd5-13862e6defd8-kube-api-access-nf5jt\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.061370 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-log-httpd\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.061700 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-run-httpd\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.067112 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-scripts\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.068127 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-config-data\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.069274 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.072439 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.080656 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf5jt\" (UniqueName: \"kubernetes.io/projected/f2746b48-57ff-4848-9fd5-13862e6defd8-kube-api-access-nf5jt\") pod \"ceilometer-0\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.209448 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.523961 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:11 crc kubenswrapper[4770]: W1004 03:23:11.534240 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2746b48_57ff_4848_9fd5_13862e6defd8.slice/crio-109e6553e420c5076789b8a09e55c5782b109d1c4fea3e61d14b2d80202a9611 WatchSource:0}: Error finding container 109e6553e420c5076789b8a09e55c5782b109d1c4fea3e61d14b2d80202a9611: Status 404 returned error can't find the container with id 109e6553e420c5076789b8a09e55c5782b109d1c4fea3e61d14b2d80202a9611 Oct 04 03:23:11 crc kubenswrapper[4770]: I1004 03:23:11.687819 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be265de0-637c-43e8-ab8d-60eff6e1d55b" path="/var/lib/kubelet/pods/be265de0-637c-43e8-ab8d-60eff6e1d55b/volumes" Oct 04 03:23:12 crc kubenswrapper[4770]: I1004 03:23:12.532589 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerStarted","Data":"65e63dbb5b0c9b461f2c4bfb021835089675895e7de625715395587714b49d15"} Oct 04 03:23:12 crc kubenswrapper[4770]: I1004 03:23:12.533105 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerStarted","Data":"109e6553e420c5076789b8a09e55c5782b109d1c4fea3e61d14b2d80202a9611"} Oct 04 03:23:17 crc kubenswrapper[4770]: I1004 03:23:17.601690 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerStarted","Data":"aeb1436e6d02983ff4efd869dcc5b939283d4b9bf4c8232806b9bf33f990a1c6"} Oct 04 03:23:17 crc kubenswrapper[4770]: I1004 03:23:17.604789 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pz6zp" event={"ID":"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8","Type":"ContainerStarted","Data":"715a0efbdfc6a5cbb344ca18f4856780c23bb1adeabccdc5c8eda7ca6149c644"} Oct 04 03:23:17 crc kubenswrapper[4770]: I1004 03:23:17.638395 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-pz6zp" podStartSLOduration=2.33681026 podStartE2EDuration="9.638310674s" podCreationTimestamp="2025-10-04 03:23:08 +0000 UTC" firstStartedPulling="2025-10-04 03:23:09.597146649 +0000 UTC m=+1200.889156371" lastFinishedPulling="2025-10-04 03:23:16.898647073 +0000 UTC m=+1208.190656785" observedRunningTime="2025-10-04 03:23:17.63013046 +0000 UTC m=+1208.922140182" watchObservedRunningTime="2025-10-04 03:23:17.638310674 +0000 UTC m=+1208.930320416" Oct 04 03:23:18 crc kubenswrapper[4770]: I1004 03:23:18.624654 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerStarted","Data":"3944da16b53c7e9a4a5e1b302b1d3a988d3664cb04d82d3a345d8264c57ae1f9"} Oct 04 03:23:19 crc kubenswrapper[4770]: I1004 03:23:19.657713 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerStarted","Data":"ea5b04f24ed7e7ce50dead47790400dc0ff130ec57ae8f47fe60ecd17e750aa1"} Oct 04 03:23:19 crc kubenswrapper[4770]: I1004 03:23:19.658446 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 03:23:19 crc kubenswrapper[4770]: I1004 03:23:19.703206 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.2739377 podStartE2EDuration="9.703176496s" podCreationTimestamp="2025-10-04 03:23:10 +0000 UTC" firstStartedPulling="2025-10-04 03:23:11.538193861 +0000 UTC m=+1202.830203583" lastFinishedPulling="2025-10-04 03:23:18.967432667 +0000 UTC m=+1210.259442379" observedRunningTime="2025-10-04 03:23:19.682617338 +0000 UTC m=+1210.974627060" watchObservedRunningTime="2025-10-04 03:23:19.703176496 +0000 UTC m=+1210.995186208" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.027301 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.028617 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="sg-core" containerID="cri-o://3944da16b53c7e9a4a5e1b302b1d3a988d3664cb04d82d3a345d8264c57ae1f9" gracePeriod=30 Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.028642 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="ceilometer-notification-agent" containerID="cri-o://aeb1436e6d02983ff4efd869dcc5b939283d4b9bf4c8232806b9bf33f990a1c6" gracePeriod=30 Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.028716 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="proxy-httpd" containerID="cri-o://ea5b04f24ed7e7ce50dead47790400dc0ff130ec57ae8f47fe60ecd17e750aa1" gracePeriod=30 Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.028812 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="ceilometer-central-agent" containerID="cri-o://65e63dbb5b0c9b461f2c4bfb021835089675895e7de625715395587714b49d15" gracePeriod=30 Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.696438 4770 generic.go:334] "Generic (PLEG): container finished" podID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerID="ea5b04f24ed7e7ce50dead47790400dc0ff130ec57ae8f47fe60ecd17e750aa1" exitCode=0 Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.696915 4770 generic.go:334] "Generic (PLEG): container finished" podID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerID="3944da16b53c7e9a4a5e1b302b1d3a988d3664cb04d82d3a345d8264c57ae1f9" exitCode=2 Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.696927 4770 generic.go:334] "Generic (PLEG): container finished" podID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerID="aeb1436e6d02983ff4efd869dcc5b939283d4b9bf4c8232806b9bf33f990a1c6" exitCode=0 Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.696938 4770 generic.go:334] "Generic (PLEG): container finished" podID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerID="65e63dbb5b0c9b461f2c4bfb021835089675895e7de625715395587714b49d15" exitCode=0 Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.696527 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerDied","Data":"ea5b04f24ed7e7ce50dead47790400dc0ff130ec57ae8f47fe60ecd17e750aa1"} Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.697040 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerDied","Data":"3944da16b53c7e9a4a5e1b302b1d3a988d3664cb04d82d3a345d8264c57ae1f9"} Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.697064 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerDied","Data":"aeb1436e6d02983ff4efd869dcc5b939283d4b9bf4c8232806b9bf33f990a1c6"} Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.697077 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerDied","Data":"65e63dbb5b0c9b461f2c4bfb021835089675895e7de625715395587714b49d15"} Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.697088 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f2746b48-57ff-4848-9fd5-13862e6defd8","Type":"ContainerDied","Data":"109e6553e420c5076789b8a09e55c5782b109d1c4fea3e61d14b2d80202a9611"} Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.697122 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="109e6553e420c5076789b8a09e55c5782b109d1c4fea3e61d14b2d80202a9611" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.732306 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.825897 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf5jt\" (UniqueName: \"kubernetes.io/projected/f2746b48-57ff-4848-9fd5-13862e6defd8-kube-api-access-nf5jt\") pod \"f2746b48-57ff-4848-9fd5-13862e6defd8\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.825952 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-log-httpd\") pod \"f2746b48-57ff-4848-9fd5-13862e6defd8\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.825993 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-scripts\") pod \"f2746b48-57ff-4848-9fd5-13862e6defd8\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.826087 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-run-httpd\") pod \"f2746b48-57ff-4848-9fd5-13862e6defd8\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.826104 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-config-data\") pod \"f2746b48-57ff-4848-9fd5-13862e6defd8\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.826142 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-sg-core-conf-yaml\") pod \"f2746b48-57ff-4848-9fd5-13862e6defd8\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.826233 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-combined-ca-bundle\") pod \"f2746b48-57ff-4848-9fd5-13862e6defd8\" (UID: \"f2746b48-57ff-4848-9fd5-13862e6defd8\") " Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.828543 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f2746b48-57ff-4848-9fd5-13862e6defd8" (UID: "f2746b48-57ff-4848-9fd5-13862e6defd8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.828764 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f2746b48-57ff-4848-9fd5-13862e6defd8" (UID: "f2746b48-57ff-4848-9fd5-13862e6defd8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.834685 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-scripts" (OuterVolumeSpecName: "scripts") pod "f2746b48-57ff-4848-9fd5-13862e6defd8" (UID: "f2746b48-57ff-4848-9fd5-13862e6defd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.834912 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2746b48-57ff-4848-9fd5-13862e6defd8-kube-api-access-nf5jt" (OuterVolumeSpecName: "kube-api-access-nf5jt") pod "f2746b48-57ff-4848-9fd5-13862e6defd8" (UID: "f2746b48-57ff-4848-9fd5-13862e6defd8"). InnerVolumeSpecName "kube-api-access-nf5jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.866435 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f2746b48-57ff-4848-9fd5-13862e6defd8" (UID: "f2746b48-57ff-4848-9fd5-13862e6defd8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.929275 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf5jt\" (UniqueName: \"kubernetes.io/projected/f2746b48-57ff-4848-9fd5-13862e6defd8-kube-api-access-nf5jt\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.929326 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.929339 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.929355 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f2746b48-57ff-4848-9fd5-13862e6defd8-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.929364 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:22 crc kubenswrapper[4770]: I1004 03:23:22.939415 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2746b48-57ff-4848-9fd5-13862e6defd8" (UID: "f2746b48-57ff-4848-9fd5-13862e6defd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.010070 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-config-data" (OuterVolumeSpecName: "config-data") pod "f2746b48-57ff-4848-9fd5-13862e6defd8" (UID: "f2746b48-57ff-4848-9fd5-13862e6defd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.030914 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.030943 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2746b48-57ff-4848-9fd5-13862e6defd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.705668 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.746140 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.755279 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.776139 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:23 crc kubenswrapper[4770]: E1004 03:23:23.776587 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="ceilometer-central-agent" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.776605 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="ceilometer-central-agent" Oct 04 03:23:23 crc kubenswrapper[4770]: E1004 03:23:23.776617 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="ceilometer-notification-agent" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.776625 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="ceilometer-notification-agent" Oct 04 03:23:23 crc kubenswrapper[4770]: E1004 03:23:23.776636 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="proxy-httpd" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.776643 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="proxy-httpd" Oct 04 03:23:23 crc kubenswrapper[4770]: E1004 03:23:23.776683 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="sg-core" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.776689 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="sg-core" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.776882 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="sg-core" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.776890 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="proxy-httpd" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.776898 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="ceilometer-notification-agent" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.776928 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" containerName="ceilometer-central-agent" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.778562 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.782341 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.783577 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.784240 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.848094 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-log-httpd\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.848147 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-scripts\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.848215 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-config-data\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.848241 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.848664 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.848824 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-run-httpd\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.848866 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk9jh\" (UniqueName: \"kubernetes.io/projected/f4edadec-fa14-45b6-9a4f-83c328b8e962-kube-api-access-qk9jh\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.950395 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-run-httpd\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.950826 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk9jh\" (UniqueName: \"kubernetes.io/projected/f4edadec-fa14-45b6-9a4f-83c328b8e962-kube-api-access-qk9jh\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.950891 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-log-httpd\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.950925 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-scripts\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.950986 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-config-data\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.951053 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.951143 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.952397 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-log-httpd\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.952655 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-run-httpd\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.956585 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.960712 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-config-data\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.961399 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.962549 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-scripts\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:23 crc kubenswrapper[4770]: I1004 03:23:23.976214 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk9jh\" (UniqueName: \"kubernetes.io/projected/f4edadec-fa14-45b6-9a4f-83c328b8e962-kube-api-access-qk9jh\") pod \"ceilometer-0\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " pod="openstack/ceilometer-0" Oct 04 03:23:24 crc kubenswrapper[4770]: I1004 03:23:24.098307 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:23:24 crc kubenswrapper[4770]: W1004 03:23:24.644721 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4edadec_fa14_45b6_9a4f_83c328b8e962.slice/crio-1302a41daa208af21523723ba6416b9d9372becea8f9071fbe29343ef2de2878 WatchSource:0}: Error finding container 1302a41daa208af21523723ba6416b9d9372becea8f9071fbe29343ef2de2878: Status 404 returned error can't find the container with id 1302a41daa208af21523723ba6416b9d9372becea8f9071fbe29343ef2de2878 Oct 04 03:23:24 crc kubenswrapper[4770]: I1004 03:23:24.644958 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:23:24 crc kubenswrapper[4770]: I1004 03:23:24.717186 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerStarted","Data":"1302a41daa208af21523723ba6416b9d9372becea8f9071fbe29343ef2de2878"} Oct 04 03:23:25 crc kubenswrapper[4770]: I1004 03:23:25.692217 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2746b48-57ff-4848-9fd5-13862e6defd8" path="/var/lib/kubelet/pods/f2746b48-57ff-4848-9fd5-13862e6defd8/volumes" Oct 04 03:23:25 crc kubenswrapper[4770]: I1004 03:23:25.729642 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerStarted","Data":"ef21a15182a9e6690eea104b5b81ffb153a9e79e9b7fdc87b31ad749a9ad7a7e"} Oct 04 03:23:26 crc kubenswrapper[4770]: I1004 03:23:26.737594 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerStarted","Data":"94d10694f73f3942e76d314307eef0e1c606808ce5d537b5bc88766b3a8aece8"} Oct 04 03:23:27 crc kubenswrapper[4770]: I1004 03:23:27.750580 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerStarted","Data":"01d04621926a93fa567a19a32594e917816c2218b6b97c3b6209f6220ded3c3d"} Oct 04 03:23:28 crc kubenswrapper[4770]: I1004 03:23:28.767050 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerStarted","Data":"3d9b2277d89aba3b4b5d07d07f7efef56ff6e822cdc71037566424b51ef55257"} Oct 04 03:23:28 crc kubenswrapper[4770]: I1004 03:23:28.770036 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 03:23:28 crc kubenswrapper[4770]: I1004 03:23:28.775655 4770 generic.go:334] "Generic (PLEG): container finished" podID="acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8" containerID="715a0efbdfc6a5cbb344ca18f4856780c23bb1adeabccdc5c8eda7ca6149c644" exitCode=0 Oct 04 03:23:28 crc kubenswrapper[4770]: I1004 03:23:28.775750 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pz6zp" event={"ID":"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8","Type":"ContainerDied","Data":"715a0efbdfc6a5cbb344ca18f4856780c23bb1adeabccdc5c8eda7ca6149c644"} Oct 04 03:23:28 crc kubenswrapper[4770]: I1004 03:23:28.816625 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.3265031560000002 podStartE2EDuration="5.816602314s" podCreationTimestamp="2025-10-04 03:23:23 +0000 UTC" firstStartedPulling="2025-10-04 03:23:24.650971172 +0000 UTC m=+1215.942980924" lastFinishedPulling="2025-10-04 03:23:28.14107034 +0000 UTC m=+1219.433080082" observedRunningTime="2025-10-04 03:23:28.812117187 +0000 UTC m=+1220.104126919" watchObservedRunningTime="2025-10-04 03:23:28.816602314 +0000 UTC m=+1220.108612056" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.219814 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.327862 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6n69\" (UniqueName: \"kubernetes.io/projected/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-kube-api-access-f6n69\") pod \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.327949 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-combined-ca-bundle\") pod \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.327995 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-scripts\") pod \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.328058 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-config-data\") pod \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\" (UID: \"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8\") " Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.341303 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-scripts" (OuterVolumeSpecName: "scripts") pod "acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8" (UID: "acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.341329 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-kube-api-access-f6n69" (OuterVolumeSpecName: "kube-api-access-f6n69") pod "acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8" (UID: "acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8"). InnerVolumeSpecName "kube-api-access-f6n69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.361274 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8" (UID: "acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.361959 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-config-data" (OuterVolumeSpecName: "config-data") pod "acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8" (UID: "acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.430318 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.430348 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.430358 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6n69\" (UniqueName: \"kubernetes.io/projected/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-kube-api-access-f6n69\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.430368 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.799803 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pz6zp" event={"ID":"acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8","Type":"ContainerDied","Data":"f9863f48804111d21fb32b80cc6a74de21f89da4257c2a405dcd9ca7072c39c5"} Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.799856 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9863f48804111d21fb32b80cc6a74de21f89da4257c2a405dcd9ca7072c39c5" Oct 04 03:23:30 crc kubenswrapper[4770]: I1004 03:23:30.799933 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pz6zp" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.017985 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 03:23:31 crc kubenswrapper[4770]: E1004 03:23:31.021259 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8" containerName="nova-cell0-conductor-db-sync" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.021289 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8" containerName="nova-cell0-conductor-db-sync" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.021578 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8" containerName="nova-cell0-conductor-db-sync" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.022322 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.026235 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.026671 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hdwhp" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.040401 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.145094 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw9wr\" (UniqueName: \"kubernetes.io/projected/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-kube-api-access-xw9wr\") pod \"nova-cell0-conductor-0\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.145546 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.145593 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.247600 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.247677 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.247809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw9wr\" (UniqueName: \"kubernetes.io/projected/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-kube-api-access-xw9wr\") pod \"nova-cell0-conductor-0\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.258024 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.272034 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw9wr\" (UniqueName: \"kubernetes.io/projected/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-kube-api-access-xw9wr\") pod \"nova-cell0-conductor-0\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.272294 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.381380 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.795648 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.796054 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:23:31 crc kubenswrapper[4770]: I1004 03:23:31.853637 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 03:23:32 crc kubenswrapper[4770]: I1004 03:23:32.824042 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e826d9bb-183e-401b-bc8b-6cfe0938a7f6","Type":"ContainerStarted","Data":"aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092"} Oct 04 03:23:32 crc kubenswrapper[4770]: I1004 03:23:32.824624 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e826d9bb-183e-401b-bc8b-6cfe0938a7f6","Type":"ContainerStarted","Data":"52a6d8bf39ba27e9c73e901e7a0cf29da2dad4207a79420d7a6b35d4610069d1"} Oct 04 03:23:32 crc kubenswrapper[4770]: I1004 03:23:32.825617 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:32 crc kubenswrapper[4770]: I1004 03:23:32.864178 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.864150129 podStartE2EDuration="2.864150129s" podCreationTimestamp="2025-10-04 03:23:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:23:32.852194446 +0000 UTC m=+1224.144204198" watchObservedRunningTime="2025-10-04 03:23:32.864150129 +0000 UTC m=+1224.156159881" Oct 04 03:23:41 crc kubenswrapper[4770]: I1004 03:23:41.427516 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.136423 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-kgblq"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.137864 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.140529 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.141039 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.157124 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kgblq"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.280846 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-config-data\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.280991 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.281117 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-scripts\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.281161 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4trrk\" (UniqueName: \"kubernetes.io/projected/e6129f31-e510-495e-a4ad-2976f2054635-kube-api-access-4trrk\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.384038 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-config-data\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.384767 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.384832 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-scripts\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.384902 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4trrk\" (UniqueName: \"kubernetes.io/projected/e6129f31-e510-495e-a4ad-2976f2054635-kube-api-access-4trrk\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.396580 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-scripts\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.398686 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-config-data\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.401524 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.408894 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.410713 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.413051 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4trrk\" (UniqueName: \"kubernetes.io/projected/e6129f31-e510-495e-a4ad-2976f2054635-kube-api-access-4trrk\") pod \"nova-cell0-cell-mapping-kgblq\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.424519 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.424750 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.426773 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.429270 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.438327 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.445055 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.446957 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.455436 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.461462 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.487801 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.487872 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.487967 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70ea433c-10b4-4fbf-a377-9af4cafd9584-logs\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.487996 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.488083 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-config-data\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.488123 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpfxf\" (UniqueName: \"kubernetes.io/projected/e79627e4-ecd2-48f3-90c2-64c68bf495f9-kube-api-access-wpfxf\") pod \"nova-cell1-novncproxy-0\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.488151 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5lds\" (UniqueName: \"kubernetes.io/projected/70ea433c-10b4-4fbf-a377-9af4cafd9584-kube-api-access-f5lds\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.493866 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.512706 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589307 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589353 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589411 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589440 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70ea433c-10b4-4fbf-a377-9af4cafd9584-logs\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589458 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589503 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-config-data\") pod \"nova-scheduler-0\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589523 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t56ln\" (UniqueName: \"kubernetes.io/projected/5ca2b7ef-1d22-4377-8a7d-272d603e0542-kube-api-access-t56ln\") pod \"nova-scheduler-0\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589560 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-config-data\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589592 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpfxf\" (UniqueName: \"kubernetes.io/projected/e79627e4-ecd2-48f3-90c2-64c68bf495f9-kube-api-access-wpfxf\") pod \"nova-cell1-novncproxy-0\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.589627 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5lds\" (UniqueName: \"kubernetes.io/projected/70ea433c-10b4-4fbf-a377-9af4cafd9584-kube-api-access-f5lds\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.592610 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70ea433c-10b4-4fbf-a377-9af4cafd9584-logs\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.601556 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-config-data\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.603701 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.608794 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.627767 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.640991 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpfxf\" (UniqueName: \"kubernetes.io/projected/e79627e4-ecd2-48f3-90c2-64c68bf495f9-kube-api-access-wpfxf\") pod \"nova-cell1-novncproxy-0\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.650140 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5lds\" (UniqueName: \"kubernetes.io/projected/70ea433c-10b4-4fbf-a377-9af4cafd9584-kube-api-access-f5lds\") pod \"nova-api-0\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.667709 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.669940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.689749 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.694763 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-config-data\") pod \"nova-scheduler-0\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.694814 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t56ln\" (UniqueName: \"kubernetes.io/projected/5ca2b7ef-1d22-4377-8a7d-272d603e0542-kube-api-access-t56ln\") pod \"nova-scheduler-0\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.695144 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.695621 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.705761 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-config-data\") pod \"nova-scheduler-0\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.715889 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.717105 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t56ln\" (UniqueName: \"kubernetes.io/projected/5ca2b7ef-1d22-4377-8a7d-272d603e0542-kube-api-access-t56ln\") pod \"nova-scheduler-0\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.752337 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-47pc5"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.753750 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.770413 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-47pc5"] Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.796972 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-config-data\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.797529 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5177412-3084-4b06-92cd-6d09f56b8b9d-logs\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.797560 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzcpp\" (UniqueName: \"kubernetes.io/projected/b5177412-3084-4b06-92cd-6d09f56b8b9d-kube-api-access-pzcpp\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.797737 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.838184 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.851162 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.865758 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.899814 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.909471 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq8sg\" (UniqueName: \"kubernetes.io/projected/11acd15b-b791-4869-a16b-37d6030b2079-kube-api-access-lq8sg\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.909832 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.909879 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-svc\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.910053 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-config\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.910145 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-config-data\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.910175 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5177412-3084-4b06-92cd-6d09f56b8b9d-logs\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.910249 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzcpp\" (UniqueName: \"kubernetes.io/projected/b5177412-3084-4b06-92cd-6d09f56b8b9d-kube-api-access-pzcpp\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.910292 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.910447 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-swift-storage-0\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.921122 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5177412-3084-4b06-92cd-6d09f56b8b9d-logs\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.925276 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-config-data\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.943206 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:42 crc kubenswrapper[4770]: I1004 03:23:42.966466 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzcpp\" (UniqueName: \"kubernetes.io/projected/b5177412-3084-4b06-92cd-6d09f56b8b9d-kube-api-access-pzcpp\") pod \"nova-metadata-0\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " pod="openstack/nova-metadata-0" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.015826 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq8sg\" (UniqueName: \"kubernetes.io/projected/11acd15b-b791-4869-a16b-37d6030b2079-kube-api-access-lq8sg\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.015892 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.015912 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-svc\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.015960 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-config\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.016037 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.016083 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-swift-storage-0\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.017333 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-swift-storage-0\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.018363 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.021523 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-svc\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.022064 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-config\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.022544 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.053199 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq8sg\" (UniqueName: \"kubernetes.io/projected/11acd15b-b791-4869-a16b-37d6030b2079-kube-api-access-lq8sg\") pod \"dnsmasq-dns-7d6d46f6cf-47pc5\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.105431 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.117263 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.325832 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcntx"] Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.327628 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.330812 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.330953 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.344125 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcntx"] Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.425525 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dgbb\" (UniqueName: \"kubernetes.io/projected/670c7093-0b8d-45d2-91f1-fccd2f275adf-kube-api-access-6dgbb\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.425587 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-scripts\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.425619 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-config-data\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.425643 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.504436 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kgblq"] Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.528946 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.529169 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dgbb\" (UniqueName: \"kubernetes.io/projected/670c7093-0b8d-45d2-91f1-fccd2f275adf-kube-api-access-6dgbb\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.529306 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-scripts\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.529348 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-config-data\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.545246 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.546935 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-config-data\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.571737 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-scripts\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.622881 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dgbb\" (UniqueName: \"kubernetes.io/projected/670c7093-0b8d-45d2-91f1-fccd2f275adf-kube-api-access-6dgbb\") pod \"nova-cell1-conductor-db-sync-rcntx\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.657247 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.714034 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.726252 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:23:43 crc kubenswrapper[4770]: W1004 03:23:43.747926 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ca2b7ef_1d22_4377_8a7d_272d603e0542.slice/crio-9db9c3d6d964c911ab9ae9ccf2cf5c734450e47a806bb805cddf02a1c8a78571 WatchSource:0}: Error finding container 9db9c3d6d964c911ab9ae9ccf2cf5c734450e47a806bb805cddf02a1c8a78571: Status 404 returned error can't find the container with id 9db9c3d6d964c911ab9ae9ccf2cf5c734450e47a806bb805cddf02a1c8a78571 Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.879679 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:43 crc kubenswrapper[4770]: W1004 03:23:43.880745 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5177412_3084_4b06_92cd_6d09f56b8b9d.slice/crio-ba2ee8a0c41de0c59641fb216ba21f010b3873649be0d23528fdfe84fbc4a19f WatchSource:0}: Error finding container ba2ee8a0c41de0c59641fb216ba21f010b3873649be0d23528fdfe84fbc4a19f: Status 404 returned error can't find the container with id ba2ee8a0c41de0c59641fb216ba21f010b3873649be0d23528fdfe84fbc4a19f Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.907819 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:23:43 crc kubenswrapper[4770]: W1004 03:23:43.994688 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11acd15b_b791_4869_a16b_37d6030b2079.slice/crio-0c3c968c83795ac1a7b40ac32ed075d700bd636eb4e9a6d04a4f18d663018271 WatchSource:0}: Error finding container 0c3c968c83795ac1a7b40ac32ed075d700bd636eb4e9a6d04a4f18d663018271: Status 404 returned error can't find the container with id 0c3c968c83795ac1a7b40ac32ed075d700bd636eb4e9a6d04a4f18d663018271 Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.997058 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-47pc5"] Oct 04 03:23:43 crc kubenswrapper[4770]: I1004 03:23:43.997308 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5177412-3084-4b06-92cd-6d09f56b8b9d","Type":"ContainerStarted","Data":"ba2ee8a0c41de0c59641fb216ba21f010b3873649be0d23528fdfe84fbc4a19f"} Oct 04 03:23:44 crc kubenswrapper[4770]: I1004 03:23:44.001899 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kgblq" event={"ID":"e6129f31-e510-495e-a4ad-2976f2054635","Type":"ContainerStarted","Data":"13d96ab8d0b2b2495f205c846512632387d84f94e00627632da603ce6493db15"} Oct 04 03:23:44 crc kubenswrapper[4770]: I1004 03:23:44.007645 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5ca2b7ef-1d22-4377-8a7d-272d603e0542","Type":"ContainerStarted","Data":"9db9c3d6d964c911ab9ae9ccf2cf5c734450e47a806bb805cddf02a1c8a78571"} Oct 04 03:23:44 crc kubenswrapper[4770]: I1004 03:23:44.016035 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70ea433c-10b4-4fbf-a377-9af4cafd9584","Type":"ContainerStarted","Data":"dbac0a868aaa79430f6b01563cab67a422e7cbfb4a16949dc4625d59b863e827"} Oct 04 03:23:44 crc kubenswrapper[4770]: I1004 03:23:44.021712 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e79627e4-ecd2-48f3-90c2-64c68bf495f9","Type":"ContainerStarted","Data":"679280c5e7ced30a83dd00bc0cd88ddd106f6d9b6be82969897f7d878c17d052"} Oct 04 03:23:44 crc kubenswrapper[4770]: W1004 03:23:44.224729 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod670c7093_0b8d_45d2_91f1_fccd2f275adf.slice/crio-f40aab7577ba5dc358f8364482ea5dae897391faaa066ae47331200e4223941d WatchSource:0}: Error finding container f40aab7577ba5dc358f8364482ea5dae897391faaa066ae47331200e4223941d: Status 404 returned error can't find the container with id f40aab7577ba5dc358f8364482ea5dae897391faaa066ae47331200e4223941d Oct 04 03:23:44 crc kubenswrapper[4770]: I1004 03:23:44.230503 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcntx"] Oct 04 03:23:45 crc kubenswrapper[4770]: I1004 03:23:45.034637 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rcntx" event={"ID":"670c7093-0b8d-45d2-91f1-fccd2f275adf","Type":"ContainerStarted","Data":"a24fe25e59f6e2eaf70b6a4057291cd447499c578bdba6d5c9b4583baff28184"} Oct 04 03:23:45 crc kubenswrapper[4770]: I1004 03:23:45.035331 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rcntx" event={"ID":"670c7093-0b8d-45d2-91f1-fccd2f275adf","Type":"ContainerStarted","Data":"f40aab7577ba5dc358f8364482ea5dae897391faaa066ae47331200e4223941d"} Oct 04 03:23:45 crc kubenswrapper[4770]: I1004 03:23:45.038513 4770 generic.go:334] "Generic (PLEG): container finished" podID="11acd15b-b791-4869-a16b-37d6030b2079" containerID="5a21c9e7067b28271f09914a49ea21aa7fa3dfb532c1ba2e54ec769996716b77" exitCode=0 Oct 04 03:23:45 crc kubenswrapper[4770]: I1004 03:23:45.038613 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" event={"ID":"11acd15b-b791-4869-a16b-37d6030b2079","Type":"ContainerDied","Data":"5a21c9e7067b28271f09914a49ea21aa7fa3dfb532c1ba2e54ec769996716b77"} Oct 04 03:23:45 crc kubenswrapper[4770]: I1004 03:23:45.038649 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" event={"ID":"11acd15b-b791-4869-a16b-37d6030b2079","Type":"ContainerStarted","Data":"0c3c968c83795ac1a7b40ac32ed075d700bd636eb4e9a6d04a4f18d663018271"} Oct 04 03:23:45 crc kubenswrapper[4770]: I1004 03:23:45.043560 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kgblq" event={"ID":"e6129f31-e510-495e-a4ad-2976f2054635","Type":"ContainerStarted","Data":"901e11a4d982f87671e376de02bae186efd672322f8829ca222a35929dc8ff60"} Oct 04 03:23:45 crc kubenswrapper[4770]: I1004 03:23:45.062550 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-rcntx" podStartSLOduration=2.062513573 podStartE2EDuration="2.062513573s" podCreationTimestamp="2025-10-04 03:23:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:23:45.058382655 +0000 UTC m=+1236.350392367" watchObservedRunningTime="2025-10-04 03:23:45.062513573 +0000 UTC m=+1236.354523285" Oct 04 03:23:45 crc kubenswrapper[4770]: I1004 03:23:45.091282 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-kgblq" podStartSLOduration=3.091260665 podStartE2EDuration="3.091260665s" podCreationTimestamp="2025-10-04 03:23:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:23:45.072656758 +0000 UTC m=+1236.364666490" watchObservedRunningTime="2025-10-04 03:23:45.091260665 +0000 UTC m=+1236.383270377" Oct 04 03:23:46 crc kubenswrapper[4770]: I1004 03:23:46.369151 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:46 crc kubenswrapper[4770]: I1004 03:23:46.391810 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.112528 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5177412-3084-4b06-92cd-6d09f56b8b9d","Type":"ContainerStarted","Data":"a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707"} Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.113641 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5177412-3084-4b06-92cd-6d09f56b8b9d","Type":"ContainerStarted","Data":"8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7"} Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.112785 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerName="nova-metadata-metadata" containerID="cri-o://a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707" gracePeriod=30 Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.112722 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerName="nova-metadata-log" containerID="cri-o://8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7" gracePeriod=30 Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.154673 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" event={"ID":"11acd15b-b791-4869-a16b-37d6030b2079","Type":"ContainerStarted","Data":"8202959a2d6a9e3319d4f040fed6d702380bf5fbfc420149e5eda797c7472b95"} Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.154900 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.158716 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5ca2b7ef-1d22-4377-8a7d-272d603e0542","Type":"ContainerStarted","Data":"69c38b6bd4986969d1b588c880ada87f4e0765ac8cedeb159ee4c7a5044d6898"} Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.166523 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70ea433c-10b4-4fbf-a377-9af4cafd9584","Type":"ContainerStarted","Data":"6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e"} Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.166665 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70ea433c-10b4-4fbf-a377-9af4cafd9584","Type":"ContainerStarted","Data":"2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624"} Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.180481 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e79627e4-ecd2-48f3-90c2-64c68bf495f9","Type":"ContainerStarted","Data":"de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d"} Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.180736 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e79627e4-ecd2-48f3-90c2-64c68bf495f9" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d" gracePeriod=30 Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.182235 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.312242085 podStartE2EDuration="8.182169957s" podCreationTimestamp="2025-10-04 03:23:42 +0000 UTC" firstStartedPulling="2025-10-04 03:23:43.895780656 +0000 UTC m=+1235.187790368" lastFinishedPulling="2025-10-04 03:23:48.765708517 +0000 UTC m=+1240.057718240" observedRunningTime="2025-10-04 03:23:50.145800146 +0000 UTC m=+1241.437809868" watchObservedRunningTime="2025-10-04 03:23:50.182169957 +0000 UTC m=+1241.474179689" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.207329 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" podStartSLOduration=8.207298994 podStartE2EDuration="8.207298994s" podCreationTimestamp="2025-10-04 03:23:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:23:50.197132309 +0000 UTC m=+1241.489142051" watchObservedRunningTime="2025-10-04 03:23:50.207298994 +0000 UTC m=+1241.499308736" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.219196 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.196171369 podStartE2EDuration="8.219164335s" podCreationTimestamp="2025-10-04 03:23:42 +0000 UTC" firstStartedPulling="2025-10-04 03:23:43.742691741 +0000 UTC m=+1235.034701453" lastFinishedPulling="2025-10-04 03:23:48.765684707 +0000 UTC m=+1240.057694419" observedRunningTime="2025-10-04 03:23:50.211127495 +0000 UTC m=+1241.503137237" watchObservedRunningTime="2025-10-04 03:23:50.219164335 +0000 UTC m=+1241.511174057" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.251142 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.3960689090000002 podStartE2EDuration="8.251107771s" podCreationTimestamp="2025-10-04 03:23:42 +0000 UTC" firstStartedPulling="2025-10-04 03:23:43.919802174 +0000 UTC m=+1235.211811886" lastFinishedPulling="2025-10-04 03:23:48.774841036 +0000 UTC m=+1240.066850748" observedRunningTime="2025-10-04 03:23:50.230772909 +0000 UTC m=+1241.522782671" watchObservedRunningTime="2025-10-04 03:23:50.251107771 +0000 UTC m=+1241.543117483" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.268801 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.257554964 podStartE2EDuration="8.268766882s" podCreationTimestamp="2025-10-04 03:23:42 +0000 UTC" firstStartedPulling="2025-10-04 03:23:43.750766222 +0000 UTC m=+1235.042775924" lastFinishedPulling="2025-10-04 03:23:48.76197813 +0000 UTC m=+1240.053987842" observedRunningTime="2025-10-04 03:23:50.248285777 +0000 UTC m=+1241.540295479" watchObservedRunningTime="2025-10-04 03:23:50.268766882 +0000 UTC m=+1241.560776594" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.707416 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.835633 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-combined-ca-bundle\") pod \"b5177412-3084-4b06-92cd-6d09f56b8b9d\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.835810 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5177412-3084-4b06-92cd-6d09f56b8b9d-logs\") pod \"b5177412-3084-4b06-92cd-6d09f56b8b9d\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.836047 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzcpp\" (UniqueName: \"kubernetes.io/projected/b5177412-3084-4b06-92cd-6d09f56b8b9d-kube-api-access-pzcpp\") pod \"b5177412-3084-4b06-92cd-6d09f56b8b9d\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.836086 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-config-data\") pod \"b5177412-3084-4b06-92cd-6d09f56b8b9d\" (UID: \"b5177412-3084-4b06-92cd-6d09f56b8b9d\") " Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.837850 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5177412-3084-4b06-92cd-6d09f56b8b9d-logs" (OuterVolumeSpecName: "logs") pod "b5177412-3084-4b06-92cd-6d09f56b8b9d" (UID: "b5177412-3084-4b06-92cd-6d09f56b8b9d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.846379 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5177412-3084-4b06-92cd-6d09f56b8b9d-kube-api-access-pzcpp" (OuterVolumeSpecName: "kube-api-access-pzcpp") pod "b5177412-3084-4b06-92cd-6d09f56b8b9d" (UID: "b5177412-3084-4b06-92cd-6d09f56b8b9d"). InnerVolumeSpecName "kube-api-access-pzcpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.867306 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-config-data" (OuterVolumeSpecName: "config-data") pod "b5177412-3084-4b06-92cd-6d09f56b8b9d" (UID: "b5177412-3084-4b06-92cd-6d09f56b8b9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.888160 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5177412-3084-4b06-92cd-6d09f56b8b9d" (UID: "b5177412-3084-4b06-92cd-6d09f56b8b9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.938583 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.938631 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5177412-3084-4b06-92cd-6d09f56b8b9d-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.938647 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzcpp\" (UniqueName: \"kubernetes.io/projected/b5177412-3084-4b06-92cd-6d09f56b8b9d-kube-api-access-pzcpp\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:50 crc kubenswrapper[4770]: I1004 03:23:50.938662 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5177412-3084-4b06-92cd-6d09f56b8b9d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.196275 4770 generic.go:334] "Generic (PLEG): container finished" podID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerID="a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707" exitCode=0 Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.196716 4770 generic.go:334] "Generic (PLEG): container finished" podID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerID="8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7" exitCode=143 Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.196493 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.196360 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5177412-3084-4b06-92cd-6d09f56b8b9d","Type":"ContainerDied","Data":"a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707"} Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.196901 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5177412-3084-4b06-92cd-6d09f56b8b9d","Type":"ContainerDied","Data":"8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7"} Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.196930 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5177412-3084-4b06-92cd-6d09f56b8b9d","Type":"ContainerDied","Data":"ba2ee8a0c41de0c59641fb216ba21f010b3873649be0d23528fdfe84fbc4a19f"} Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.196959 4770 scope.go:117] "RemoveContainer" containerID="a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.244952 4770 scope.go:117] "RemoveContainer" containerID="8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.245225 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.267976 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.281281 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:51 crc kubenswrapper[4770]: E1004 03:23:51.281789 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerName="nova-metadata-log" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.281808 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerName="nova-metadata-log" Oct 04 03:23:51 crc kubenswrapper[4770]: E1004 03:23:51.281844 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerName="nova-metadata-metadata" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.281852 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerName="nova-metadata-metadata" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.282104 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerName="nova-metadata-log" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.282125 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5177412-3084-4b06-92cd-6d09f56b8b9d" containerName="nova-metadata-metadata" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.283297 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.285829 4770 scope.go:117] "RemoveContainer" containerID="a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.286483 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.287076 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.292263 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:51 crc kubenswrapper[4770]: E1004 03:23:51.308391 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707\": container with ID starting with a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707 not found: ID does not exist" containerID="a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.308447 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707"} err="failed to get container status \"a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707\": rpc error: code = NotFound desc = could not find container \"a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707\": container with ID starting with a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707 not found: ID does not exist" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.308476 4770 scope.go:117] "RemoveContainer" containerID="8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7" Oct 04 03:23:51 crc kubenswrapper[4770]: E1004 03:23:51.310608 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7\": container with ID starting with 8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7 not found: ID does not exist" containerID="8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.310790 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7"} err="failed to get container status \"8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7\": rpc error: code = NotFound desc = could not find container \"8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7\": container with ID starting with 8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7 not found: ID does not exist" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.310923 4770 scope.go:117] "RemoveContainer" containerID="a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.311619 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707"} err="failed to get container status \"a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707\": rpc error: code = NotFound desc = could not find container \"a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707\": container with ID starting with a68ef9d5cc24895e6ebf700971765541ec175ccb47e701f74f299b8ec7841707 not found: ID does not exist" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.311651 4770 scope.go:117] "RemoveContainer" containerID="8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.312637 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7"} err="failed to get container status \"8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7\": rpc error: code = NotFound desc = could not find container \"8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7\": container with ID starting with 8a8986dcdc9bd1257b8e34c2e48909297b39f2921b776085a6ac3abe26f5d3e7 not found: ID does not exist" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.454271 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.454378 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-config-data\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.454453 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c681a92-7a85-44f1-82bb-0a24460a0661-logs\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.454727 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tfz2\" (UniqueName: \"kubernetes.io/projected/8c681a92-7a85-44f1-82bb-0a24460a0661-kube-api-access-6tfz2\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.454819 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.557074 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tfz2\" (UniqueName: \"kubernetes.io/projected/8c681a92-7a85-44f1-82bb-0a24460a0661-kube-api-access-6tfz2\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.557148 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.557262 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.557290 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-config-data\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.557328 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c681a92-7a85-44f1-82bb-0a24460a0661-logs\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.558031 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c681a92-7a85-44f1-82bb-0a24460a0661-logs\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.576938 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.577324 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.577467 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-config-data\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.587841 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tfz2\" (UniqueName: \"kubernetes.io/projected/8c681a92-7a85-44f1-82bb-0a24460a0661-kube-api-access-6tfz2\") pod \"nova-metadata-0\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.633637 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:51 crc kubenswrapper[4770]: I1004 03:23:51.692786 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5177412-3084-4b06-92cd-6d09f56b8b9d" path="/var/lib/kubelet/pods/b5177412-3084-4b06-92cd-6d09f56b8b9d/volumes" Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.150133 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.213633 4770 generic.go:334] "Generic (PLEG): container finished" podID="e6129f31-e510-495e-a4ad-2976f2054635" containerID="901e11a4d982f87671e376de02bae186efd672322f8829ca222a35929dc8ff60" exitCode=0 Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.213701 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kgblq" event={"ID":"e6129f31-e510-495e-a4ad-2976f2054635","Type":"ContainerDied","Data":"901e11a4d982f87671e376de02bae186efd672322f8829ca222a35929dc8ff60"} Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.216926 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c681a92-7a85-44f1-82bb-0a24460a0661","Type":"ContainerStarted","Data":"fabd32c86786c75c6b09ea56b0fe5c71f6469d08c1e9c72d86879e030f712958"} Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.839845 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.851985 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.852557 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.867074 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.867163 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 03:23:52 crc kubenswrapper[4770]: I1004 03:23:52.924113 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.234175 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c681a92-7a85-44f1-82bb-0a24460a0661","Type":"ContainerStarted","Data":"bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356"} Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.234236 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c681a92-7a85-44f1-82bb-0a24460a0661","Type":"ContainerStarted","Data":"f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81"} Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.266574 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.266543725 podStartE2EDuration="2.266543725s" podCreationTimestamp="2025-10-04 03:23:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:23:53.264546113 +0000 UTC m=+1244.556555825" watchObservedRunningTime="2025-10-04 03:23:53.266543725 +0000 UTC m=+1244.558553467" Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.288573 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.748748 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.920821 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-config-data\") pod \"e6129f31-e510-495e-a4ad-2976f2054635\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.921068 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4trrk\" (UniqueName: \"kubernetes.io/projected/e6129f31-e510-495e-a4ad-2976f2054635-kube-api-access-4trrk\") pod \"e6129f31-e510-495e-a4ad-2976f2054635\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.921154 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-combined-ca-bundle\") pod \"e6129f31-e510-495e-a4ad-2976f2054635\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.921332 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-scripts\") pod \"e6129f31-e510-495e-a4ad-2976f2054635\" (UID: \"e6129f31-e510-495e-a4ad-2976f2054635\") " Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.932381 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-scripts" (OuterVolumeSpecName: "scripts") pod "e6129f31-e510-495e-a4ad-2976f2054635" (UID: "e6129f31-e510-495e-a4ad-2976f2054635"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.934258 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6129f31-e510-495e-a4ad-2976f2054635-kube-api-access-4trrk" (OuterVolumeSpecName: "kube-api-access-4trrk") pod "e6129f31-e510-495e-a4ad-2976f2054635" (UID: "e6129f31-e510-495e-a4ad-2976f2054635"). InnerVolumeSpecName "kube-api-access-4trrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.948445 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.948529 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.964345 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-config-data" (OuterVolumeSpecName: "config-data") pod "e6129f31-e510-495e-a4ad-2976f2054635" (UID: "e6129f31-e510-495e-a4ad-2976f2054635"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:53 crc kubenswrapper[4770]: I1004 03:23:53.976189 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6129f31-e510-495e-a4ad-2976f2054635" (UID: "e6129f31-e510-495e-a4ad-2976f2054635"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.023988 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.024055 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.024069 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4trrk\" (UniqueName: \"kubernetes.io/projected/e6129f31-e510-495e-a4ad-2976f2054635-kube-api-access-4trrk\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.024082 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6129f31-e510-495e-a4ad-2976f2054635-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.157280 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.249529 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kgblq" event={"ID":"e6129f31-e510-495e-a4ad-2976f2054635","Type":"ContainerDied","Data":"13d96ab8d0b2b2495f205c846512632387d84f94e00627632da603ce6493db15"} Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.249579 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13d96ab8d0b2b2495f205c846512632387d84f94e00627632da603ce6493db15" Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.249660 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kgblq" Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.369145 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.369437 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-log" containerID="cri-o://2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624" gracePeriod=30 Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.370130 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-api" containerID="cri-o://6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e" gracePeriod=30 Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.376557 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:23:54 crc kubenswrapper[4770]: I1004 03:23:54.402077 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:55 crc kubenswrapper[4770]: I1004 03:23:55.266954 4770 generic.go:334] "Generic (PLEG): container finished" podID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerID="2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624" exitCode=143 Oct 04 03:23:55 crc kubenswrapper[4770]: I1004 03:23:55.267083 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70ea433c-10b4-4fbf-a377-9af4cafd9584","Type":"ContainerDied","Data":"2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624"} Oct 04 03:23:55 crc kubenswrapper[4770]: I1004 03:23:55.267234 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerName="nova-metadata-log" containerID="cri-o://f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81" gracePeriod=30 Oct 04 03:23:55 crc kubenswrapper[4770]: I1004 03:23:55.267280 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerName="nova-metadata-metadata" containerID="cri-o://bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356" gracePeriod=30 Oct 04 03:23:55 crc kubenswrapper[4770]: I1004 03:23:55.267414 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5ca2b7ef-1d22-4377-8a7d-272d603e0542" containerName="nova-scheduler-scheduler" containerID="cri-o://69c38b6bd4986969d1b588c880ada87f4e0765ac8cedeb159ee4c7a5044d6898" gracePeriod=30 Oct 04 03:23:55 crc kubenswrapper[4770]: I1004 03:23:55.947572 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.075980 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-combined-ca-bundle\") pod \"8c681a92-7a85-44f1-82bb-0a24460a0661\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.076081 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c681a92-7a85-44f1-82bb-0a24460a0661-logs\") pod \"8c681a92-7a85-44f1-82bb-0a24460a0661\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.076121 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-nova-metadata-tls-certs\") pod \"8c681a92-7a85-44f1-82bb-0a24460a0661\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.076192 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-config-data\") pod \"8c681a92-7a85-44f1-82bb-0a24460a0661\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.076236 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tfz2\" (UniqueName: \"kubernetes.io/projected/8c681a92-7a85-44f1-82bb-0a24460a0661-kube-api-access-6tfz2\") pod \"8c681a92-7a85-44f1-82bb-0a24460a0661\" (UID: \"8c681a92-7a85-44f1-82bb-0a24460a0661\") " Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.078336 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c681a92-7a85-44f1-82bb-0a24460a0661-logs" (OuterVolumeSpecName: "logs") pod "8c681a92-7a85-44f1-82bb-0a24460a0661" (UID: "8c681a92-7a85-44f1-82bb-0a24460a0661"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.086353 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c681a92-7a85-44f1-82bb-0a24460a0661-kube-api-access-6tfz2" (OuterVolumeSpecName: "kube-api-access-6tfz2") pod "8c681a92-7a85-44f1-82bb-0a24460a0661" (UID: "8c681a92-7a85-44f1-82bb-0a24460a0661"). InnerVolumeSpecName "kube-api-access-6tfz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.118070 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c681a92-7a85-44f1-82bb-0a24460a0661" (UID: "8c681a92-7a85-44f1-82bb-0a24460a0661"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.131514 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-config-data" (OuterVolumeSpecName: "config-data") pod "8c681a92-7a85-44f1-82bb-0a24460a0661" (UID: "8c681a92-7a85-44f1-82bb-0a24460a0661"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.151239 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8c681a92-7a85-44f1-82bb-0a24460a0661" (UID: "8c681a92-7a85-44f1-82bb-0a24460a0661"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.179072 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.179111 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c681a92-7a85-44f1-82bb-0a24460a0661-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.179124 4770 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.179138 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c681a92-7a85-44f1-82bb-0a24460a0661-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.179149 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tfz2\" (UniqueName: \"kubernetes.io/projected/8c681a92-7a85-44f1-82bb-0a24460a0661-kube-api-access-6tfz2\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.281329 4770 generic.go:334] "Generic (PLEG): container finished" podID="670c7093-0b8d-45d2-91f1-fccd2f275adf" containerID="a24fe25e59f6e2eaf70b6a4057291cd447499c578bdba6d5c9b4583baff28184" exitCode=0 Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.281425 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rcntx" event={"ID":"670c7093-0b8d-45d2-91f1-fccd2f275adf","Type":"ContainerDied","Data":"a24fe25e59f6e2eaf70b6a4057291cd447499c578bdba6d5c9b4583baff28184"} Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.284261 4770 generic.go:334] "Generic (PLEG): container finished" podID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerID="bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356" exitCode=0 Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.284299 4770 generic.go:334] "Generic (PLEG): container finished" podID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerID="f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81" exitCode=143 Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.284308 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.284324 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c681a92-7a85-44f1-82bb-0a24460a0661","Type":"ContainerDied","Data":"bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356"} Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.284359 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c681a92-7a85-44f1-82bb-0a24460a0661","Type":"ContainerDied","Data":"f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81"} Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.284375 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c681a92-7a85-44f1-82bb-0a24460a0661","Type":"ContainerDied","Data":"fabd32c86786c75c6b09ea56b0fe5c71f6469d08c1e9c72d86879e030f712958"} Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.284406 4770 scope.go:117] "RemoveContainer" containerID="bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.318234 4770 scope.go:117] "RemoveContainer" containerID="f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.336449 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.346313 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.353473 4770 scope.go:117] "RemoveContainer" containerID="bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356" Oct 04 03:23:56 crc kubenswrapper[4770]: E1004 03:23:56.357661 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356\": container with ID starting with bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356 not found: ID does not exist" containerID="bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.357945 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356"} err="failed to get container status \"bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356\": rpc error: code = NotFound desc = could not find container \"bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356\": container with ID starting with bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356 not found: ID does not exist" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.358041 4770 scope.go:117] "RemoveContainer" containerID="f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81" Oct 04 03:23:56 crc kubenswrapper[4770]: E1004 03:23:56.358364 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81\": container with ID starting with f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81 not found: ID does not exist" containerID="f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.358389 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81"} err="failed to get container status \"f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81\": rpc error: code = NotFound desc = could not find container \"f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81\": container with ID starting with f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81 not found: ID does not exist" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.358404 4770 scope.go:117] "RemoveContainer" containerID="bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.359102 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356"} err="failed to get container status \"bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356\": rpc error: code = NotFound desc = could not find container \"bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356\": container with ID starting with bc73e85532ee76e953e10d3aa1b18a557e0ae29694081027ace415b1abc53356 not found: ID does not exist" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.359153 4770 scope.go:117] "RemoveContainer" containerID="f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.359467 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81"} err="failed to get container status \"f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81\": rpc error: code = NotFound desc = could not find container \"f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81\": container with ID starting with f4cfb343cc670607ae9c31acbab5058c1e2cc5c511b04d161581587e600cba81 not found: ID does not exist" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.375044 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:56 crc kubenswrapper[4770]: E1004 03:23:56.376656 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerName="nova-metadata-log" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.376685 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerName="nova-metadata-log" Oct 04 03:23:56 crc kubenswrapper[4770]: E1004 03:23:56.376707 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerName="nova-metadata-metadata" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.376715 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerName="nova-metadata-metadata" Oct 04 03:23:56 crc kubenswrapper[4770]: E1004 03:23:56.376731 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6129f31-e510-495e-a4ad-2976f2054635" containerName="nova-manage" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.376739 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6129f31-e510-495e-a4ad-2976f2054635" containerName="nova-manage" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.377045 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6129f31-e510-495e-a4ad-2976f2054635" containerName="nova-manage" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.377071 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerName="nova-metadata-metadata" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.377104 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c681a92-7a85-44f1-82bb-0a24460a0661" containerName="nova-metadata-log" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.378602 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.383423 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.383656 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.405435 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.485307 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-config-data\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.485383 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0271b4e-6f09-49e2-9458-d495e7bf936a-logs\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.485471 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.485516 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.485667 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77l7w\" (UniqueName: \"kubernetes.io/projected/b0271b4e-6f09-49e2-9458-d495e7bf936a-kube-api-access-77l7w\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.587728 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.588102 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.588245 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77l7w\" (UniqueName: \"kubernetes.io/projected/b0271b4e-6f09-49e2-9458-d495e7bf936a-kube-api-access-77l7w\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.588279 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-config-data\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.588324 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0271b4e-6f09-49e2-9458-d495e7bf936a-logs\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.589223 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0271b4e-6f09-49e2-9458-d495e7bf936a-logs\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.594821 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.594874 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.595063 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-config-data\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.624611 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77l7w\" (UniqueName: \"kubernetes.io/projected/b0271b4e-6f09-49e2-9458-d495e7bf936a-kube-api-access-77l7w\") pod \"nova-metadata-0\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " pod="openstack/nova-metadata-0" Oct 04 03:23:56 crc kubenswrapper[4770]: I1004 03:23:56.758244 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.300848 4770 generic.go:334] "Generic (PLEG): container finished" podID="5ca2b7ef-1d22-4377-8a7d-272d603e0542" containerID="69c38b6bd4986969d1b588c880ada87f4e0765ac8cedeb159ee4c7a5044d6898" exitCode=0 Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.301412 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5ca2b7ef-1d22-4377-8a7d-272d603e0542","Type":"ContainerDied","Data":"69c38b6bd4986969d1b588c880ada87f4e0765ac8cedeb159ee4c7a5044d6898"} Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.444989 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.557141 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.690390 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c681a92-7a85-44f1-82bb-0a24460a0661" path="/var/lib/kubelet/pods/8c681a92-7a85-44f1-82bb-0a24460a0661/volumes" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.747776 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-combined-ca-bundle\") pod \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.748098 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-config-data\") pod \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.748272 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t56ln\" (UniqueName: \"kubernetes.io/projected/5ca2b7ef-1d22-4377-8a7d-272d603e0542-kube-api-access-t56ln\") pod \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\" (UID: \"5ca2b7ef-1d22-4377-8a7d-272d603e0542\") " Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.759323 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ca2b7ef-1d22-4377-8a7d-272d603e0542-kube-api-access-t56ln" (OuterVolumeSpecName: "kube-api-access-t56ln") pod "5ca2b7ef-1d22-4377-8a7d-272d603e0542" (UID: "5ca2b7ef-1d22-4377-8a7d-272d603e0542"). InnerVolumeSpecName "kube-api-access-t56ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.760207 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.843438 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-config-data" (OuterVolumeSpecName: "config-data") pod "5ca2b7ef-1d22-4377-8a7d-272d603e0542" (UID: "5ca2b7ef-1d22-4377-8a7d-272d603e0542"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.844113 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ca2b7ef-1d22-4377-8a7d-272d603e0542" (UID: "5ca2b7ef-1d22-4377-8a7d-272d603e0542"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.850505 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-combined-ca-bundle\") pod \"670c7093-0b8d-45d2-91f1-fccd2f275adf\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.850689 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dgbb\" (UniqueName: \"kubernetes.io/projected/670c7093-0b8d-45d2-91f1-fccd2f275adf-kube-api-access-6dgbb\") pod \"670c7093-0b8d-45d2-91f1-fccd2f275adf\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.850839 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-config-data\") pod \"670c7093-0b8d-45d2-91f1-fccd2f275adf\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.850938 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-scripts\") pod \"670c7093-0b8d-45d2-91f1-fccd2f275adf\" (UID: \"670c7093-0b8d-45d2-91f1-fccd2f275adf\") " Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.851559 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t56ln\" (UniqueName: \"kubernetes.io/projected/5ca2b7ef-1d22-4377-8a7d-272d603e0542-kube-api-access-t56ln\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.851583 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.851595 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca2b7ef-1d22-4377-8a7d-272d603e0542-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.861815 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/670c7093-0b8d-45d2-91f1-fccd2f275adf-kube-api-access-6dgbb" (OuterVolumeSpecName: "kube-api-access-6dgbb") pod "670c7093-0b8d-45d2-91f1-fccd2f275adf" (UID: "670c7093-0b8d-45d2-91f1-fccd2f275adf"). InnerVolumeSpecName "kube-api-access-6dgbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.871601 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-scripts" (OuterVolumeSpecName: "scripts") pod "670c7093-0b8d-45d2-91f1-fccd2f275adf" (UID: "670c7093-0b8d-45d2-91f1-fccd2f275adf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.892505 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-config-data" (OuterVolumeSpecName: "config-data") pod "670c7093-0b8d-45d2-91f1-fccd2f275adf" (UID: "670c7093-0b8d-45d2-91f1-fccd2f275adf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.907132 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "670c7093-0b8d-45d2-91f1-fccd2f275adf" (UID: "670c7093-0b8d-45d2-91f1-fccd2f275adf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.954082 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dgbb\" (UniqueName: \"kubernetes.io/projected/670c7093-0b8d-45d2-91f1-fccd2f275adf-kube-api-access-6dgbb\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.954117 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.954127 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:57 crc kubenswrapper[4770]: I1004 03:23:57.954137 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670c7093-0b8d-45d2-91f1-fccd2f275adf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.120305 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.205393 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-zfjdv"] Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.205805 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" podUID="3ac2690b-33cd-48f3-92d9-b9a0e29442e0" containerName="dnsmasq-dns" containerID="cri-o://937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd" gracePeriod=10 Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.374913 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0271b4e-6f09-49e2-9458-d495e7bf936a","Type":"ContainerStarted","Data":"b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9"} Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.375524 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0271b4e-6f09-49e2-9458-d495e7bf936a","Type":"ContainerStarted","Data":"9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee"} Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.375539 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0271b4e-6f09-49e2-9458-d495e7bf936a","Type":"ContainerStarted","Data":"6ca0e1685f3d9eb5381783c9fa1710ec56c6c9c973345194ac0da66eed338d54"} Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.396473 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rcntx" event={"ID":"670c7093-0b8d-45d2-91f1-fccd2f275adf","Type":"ContainerDied","Data":"f40aab7577ba5dc358f8364482ea5dae897391faaa066ae47331200e4223941d"} Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.396535 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f40aab7577ba5dc358f8364482ea5dae897391faaa066ae47331200e4223941d" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.396688 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rcntx" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.442950 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5ca2b7ef-1d22-4377-8a7d-272d603e0542","Type":"ContainerDied","Data":"9db9c3d6d964c911ab9ae9ccf2cf5c734450e47a806bb805cddf02a1c8a78571"} Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.443022 4770 scope.go:117] "RemoveContainer" containerID="69c38b6bd4986969d1b588c880ada87f4e0765ac8cedeb159ee4c7a5044d6898" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.443675 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.484951 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 03:23:58 crc kubenswrapper[4770]: E1004 03:23:58.485736 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670c7093-0b8d-45d2-91f1-fccd2f275adf" containerName="nova-cell1-conductor-db-sync" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.485753 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="670c7093-0b8d-45d2-91f1-fccd2f275adf" containerName="nova-cell1-conductor-db-sync" Oct 04 03:23:58 crc kubenswrapper[4770]: E1004 03:23:58.485781 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ca2b7ef-1d22-4377-8a7d-272d603e0542" containerName="nova-scheduler-scheduler" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.485788 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ca2b7ef-1d22-4377-8a7d-272d603e0542" containerName="nova-scheduler-scheduler" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.486150 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ca2b7ef-1d22-4377-8a7d-272d603e0542" containerName="nova-scheduler-scheduler" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.486178 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="670c7093-0b8d-45d2-91f1-fccd2f275adf" containerName="nova-cell1-conductor-db-sync" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.486890 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.489847 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.523402 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.523372357 podStartE2EDuration="2.523372357s" podCreationTimestamp="2025-10-04 03:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:23:58.408952244 +0000 UTC m=+1249.700961956" watchObservedRunningTime="2025-10-04 03:23:58.523372357 +0000 UTC m=+1249.815382069" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.571147 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.582399 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.593203 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.593271 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.593315 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnfqr\" (UniqueName: \"kubernetes.io/projected/ef58a910-f85e-4e14-95b5-807741097485-kube-api-access-rnfqr\") pod \"nova-cell1-conductor-0\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.602021 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.615885 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.618061 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.622589 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.633162 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.697127 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.697201 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.697230 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnfqr\" (UniqueName: \"kubernetes.io/projected/ef58a910-f85e-4e14-95b5-807741097485-kube-api-access-rnfqr\") pod \"nova-cell1-conductor-0\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.713827 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.715786 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnfqr\" (UniqueName: \"kubernetes.io/projected/ef58a910-f85e-4e14-95b5-807741097485-kube-api-access-rnfqr\") pod \"nova-cell1-conductor-0\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.716238 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.799807 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.800355 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr5qr\" (UniqueName: \"kubernetes.io/projected/5a42a6c3-db14-49a3-bc09-79315b1c0169-kube-api-access-vr5qr\") pod \"nova-scheduler-0\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.800861 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-config-data\") pod \"nova-scheduler-0\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.826886 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.903495 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.903574 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr5qr\" (UniqueName: \"kubernetes.io/projected/5a42a6c3-db14-49a3-bc09-79315b1c0169-kube-api-access-vr5qr\") pod \"nova-scheduler-0\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.903674 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-config-data\") pod \"nova-scheduler-0\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.910604 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-config-data\") pod \"nova-scheduler-0\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.911643 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.929326 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr5qr\" (UniqueName: \"kubernetes.io/projected/5a42a6c3-db14-49a3-bc09-79315b1c0169-kube-api-access-vr5qr\") pod \"nova-scheduler-0\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " pod="openstack/nova-scheduler-0" Oct 04 03:23:58 crc kubenswrapper[4770]: I1004 03:23:58.957145 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.034189 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.211961 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-nb\") pod \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.212581 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-sb\") pod \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.212644 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-config\") pod \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.213357 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwsgh\" (UniqueName: \"kubernetes.io/projected/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-kube-api-access-qwsgh\") pod \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.213454 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-swift-storage-0\") pod \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.213493 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-svc\") pod \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\" (UID: \"3ac2690b-33cd-48f3-92d9-b9a0e29442e0\") " Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.221242 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-kube-api-access-qwsgh" (OuterVolumeSpecName: "kube-api-access-qwsgh") pod "3ac2690b-33cd-48f3-92d9-b9a0e29442e0" (UID: "3ac2690b-33cd-48f3-92d9-b9a0e29442e0"). InnerVolumeSpecName "kube-api-access-qwsgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.279433 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3ac2690b-33cd-48f3-92d9-b9a0e29442e0" (UID: "3ac2690b-33cd-48f3-92d9-b9a0e29442e0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.283744 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ac2690b-33cd-48f3-92d9-b9a0e29442e0" (UID: "3ac2690b-33cd-48f3-92d9-b9a0e29442e0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.302623 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ac2690b-33cd-48f3-92d9-b9a0e29442e0" (UID: "3ac2690b-33cd-48f3-92d9-b9a0e29442e0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.306297 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.306560 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="60a374b7-7611-4f40-b0ae-a6238903e0ea" containerName="kube-state-metrics" containerID="cri-o://6633e392a5a6f2b64dfbad7667d1b0267f86fde4b17e7db1b8808041de34c634" gracePeriod=30 Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.321424 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.321483 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwsgh\" (UniqueName: \"kubernetes.io/projected/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-kube-api-access-qwsgh\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.321519 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.321534 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.322119 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ac2690b-33cd-48f3-92d9-b9a0e29442e0" (UID: "3ac2690b-33cd-48f3-92d9-b9a0e29442e0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.335813 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-config" (OuterVolumeSpecName: "config") pod "3ac2690b-33cd-48f3-92d9-b9a0e29442e0" (UID: "3ac2690b-33cd-48f3-92d9-b9a0e29442e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.402489 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.426695 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.426754 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac2690b-33cd-48f3-92d9-b9a0e29442e0-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.468364 4770 generic.go:334] "Generic (PLEG): container finished" podID="3ac2690b-33cd-48f3-92d9-b9a0e29442e0" containerID="937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd" exitCode=0 Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.468468 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" event={"ID":"3ac2690b-33cd-48f3-92d9-b9a0e29442e0","Type":"ContainerDied","Data":"937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd"} Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.468512 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" event={"ID":"3ac2690b-33cd-48f3-92d9-b9a0e29442e0","Type":"ContainerDied","Data":"e5234fe27ef714c765747f10c0313ae31c7c9a2d6045523aa1dad71830928543"} Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.468536 4770 scope.go:117] "RemoveContainer" containerID="937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.468859 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548c5895b5-zfjdv" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.485239 4770 generic.go:334] "Generic (PLEG): container finished" podID="60a374b7-7611-4f40-b0ae-a6238903e0ea" containerID="6633e392a5a6f2b64dfbad7667d1b0267f86fde4b17e7db1b8808041de34c634" exitCode=2 Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.485387 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"60a374b7-7611-4f40-b0ae-a6238903e0ea","Type":"ContainerDied","Data":"6633e392a5a6f2b64dfbad7667d1b0267f86fde4b17e7db1b8808041de34c634"} Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.492577 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ef58a910-f85e-4e14-95b5-807741097485","Type":"ContainerStarted","Data":"7078e60ebe9016e8d310ce62e5d638d0400876fabbb7b01b95c7c917d5a6fdf3"} Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.536763 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-zfjdv"] Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.550686 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-548c5895b5-zfjdv"] Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.550985 4770 scope.go:117] "RemoveContainer" containerID="d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.606720 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:23:59 crc kubenswrapper[4770]: W1004 03:23:59.619780 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a42a6c3_db14_49a3_bc09_79315b1c0169.slice/crio-4670ac3ccfba07bfb82864f98d488b29544472d4e1c09be841dd2d46f6bfd803 WatchSource:0}: Error finding container 4670ac3ccfba07bfb82864f98d488b29544472d4e1c09be841dd2d46f6bfd803: Status 404 returned error can't find the container with id 4670ac3ccfba07bfb82864f98d488b29544472d4e1c09be841dd2d46f6bfd803 Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.635770 4770 scope.go:117] "RemoveContainer" containerID="937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd" Oct 04 03:23:59 crc kubenswrapper[4770]: E1004 03:23:59.640993 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd\": container with ID starting with 937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd not found: ID does not exist" containerID="937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.641062 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd"} err="failed to get container status \"937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd\": rpc error: code = NotFound desc = could not find container \"937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd\": container with ID starting with 937dd8ef5a7aa034c34a3a72f96de2e4a9c744c6a9dfcedd15468c76195440bd not found: ID does not exist" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.641111 4770 scope.go:117] "RemoveContainer" containerID="d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d" Oct 04 03:23:59 crc kubenswrapper[4770]: E1004 03:23:59.641679 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d\": container with ID starting with d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d not found: ID does not exist" containerID="d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.641725 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d"} err="failed to get container status \"d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d\": rpc error: code = NotFound desc = could not find container \"d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d\": container with ID starting with d7eb2ecdefc54666dd647f6edefa32a2c15de4f176f3d7c560d2af01a4d1f32d not found: ID does not exist" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.706128 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ac2690b-33cd-48f3-92d9-b9a0e29442e0" path="/var/lib/kubelet/pods/3ac2690b-33cd-48f3-92d9-b9a0e29442e0/volumes" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.706823 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ca2b7ef-1d22-4377-8a7d-272d603e0542" path="/var/lib/kubelet/pods/5ca2b7ef-1d22-4377-8a7d-272d603e0542/volumes" Oct 04 03:23:59 crc kubenswrapper[4770]: I1004 03:23:59.904034 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.049958 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpm2f\" (UniqueName: \"kubernetes.io/projected/60a374b7-7611-4f40-b0ae-a6238903e0ea-kube-api-access-vpm2f\") pod \"60a374b7-7611-4f40-b0ae-a6238903e0ea\" (UID: \"60a374b7-7611-4f40-b0ae-a6238903e0ea\") " Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.060244 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60a374b7-7611-4f40-b0ae-a6238903e0ea-kube-api-access-vpm2f" (OuterVolumeSpecName: "kube-api-access-vpm2f") pod "60a374b7-7611-4f40-b0ae-a6238903e0ea" (UID: "60a374b7-7611-4f40-b0ae-a6238903e0ea"). InnerVolumeSpecName "kube-api-access-vpm2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.159229 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpm2f\" (UniqueName: \"kubernetes.io/projected/60a374b7-7611-4f40-b0ae-a6238903e0ea-kube-api-access-vpm2f\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.303988 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.365979 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-combined-ca-bundle\") pod \"70ea433c-10b4-4fbf-a377-9af4cafd9584\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.366084 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5lds\" (UniqueName: \"kubernetes.io/projected/70ea433c-10b4-4fbf-a377-9af4cafd9584-kube-api-access-f5lds\") pod \"70ea433c-10b4-4fbf-a377-9af4cafd9584\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.366159 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-config-data\") pod \"70ea433c-10b4-4fbf-a377-9af4cafd9584\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.366455 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70ea433c-10b4-4fbf-a377-9af4cafd9584-logs\") pod \"70ea433c-10b4-4fbf-a377-9af4cafd9584\" (UID: \"70ea433c-10b4-4fbf-a377-9af4cafd9584\") " Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.367448 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70ea433c-10b4-4fbf-a377-9af4cafd9584-logs" (OuterVolumeSpecName: "logs") pod "70ea433c-10b4-4fbf-a377-9af4cafd9584" (UID: "70ea433c-10b4-4fbf-a377-9af4cafd9584"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.374170 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70ea433c-10b4-4fbf-a377-9af4cafd9584-kube-api-access-f5lds" (OuterVolumeSpecName: "kube-api-access-f5lds") pod "70ea433c-10b4-4fbf-a377-9af4cafd9584" (UID: "70ea433c-10b4-4fbf-a377-9af4cafd9584"). InnerVolumeSpecName "kube-api-access-f5lds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.419098 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-config-data" (OuterVolumeSpecName: "config-data") pod "70ea433c-10b4-4fbf-a377-9af4cafd9584" (UID: "70ea433c-10b4-4fbf-a377-9af4cafd9584"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.430967 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70ea433c-10b4-4fbf-a377-9af4cafd9584" (UID: "70ea433c-10b4-4fbf-a377-9af4cafd9584"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.477065 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.477134 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5lds\" (UniqueName: \"kubernetes.io/projected/70ea433c-10b4-4fbf-a377-9af4cafd9584-kube-api-access-f5lds\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.477151 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ea433c-10b4-4fbf-a377-9af4cafd9584-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.477170 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70ea433c-10b4-4fbf-a377-9af4cafd9584-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.527349 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a42a6c3-db14-49a3-bc09-79315b1c0169","Type":"ContainerStarted","Data":"8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1"} Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.527417 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a42a6c3-db14-49a3-bc09-79315b1c0169","Type":"ContainerStarted","Data":"4670ac3ccfba07bfb82864f98d488b29544472d4e1c09be841dd2d46f6bfd803"} Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.532183 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ef58a910-f85e-4e14-95b5-807741097485","Type":"ContainerStarted","Data":"7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda"} Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.533446 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.548285 4770 generic.go:334] "Generic (PLEG): container finished" podID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerID="6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e" exitCode=0 Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.548399 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70ea433c-10b4-4fbf-a377-9af4cafd9584","Type":"ContainerDied","Data":"6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e"} Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.548441 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70ea433c-10b4-4fbf-a377-9af4cafd9584","Type":"ContainerDied","Data":"dbac0a868aaa79430f6b01563cab67a422e7cbfb4a16949dc4625d59b863e827"} Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.548471 4770 scope.go:117] "RemoveContainer" containerID="6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.548738 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.557518 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"60a374b7-7611-4f40-b0ae-a6238903e0ea","Type":"ContainerDied","Data":"5b816be7aa73f09f3b7105f982c4bde6cfd4b19cff32c9048a336134b52a5696"} Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.557733 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.558483 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.558470699 podStartE2EDuration="2.558470699s" podCreationTimestamp="2025-10-04 03:23:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:00.553103558 +0000 UTC m=+1251.845113270" watchObservedRunningTime="2025-10-04 03:24:00.558470699 +0000 UTC m=+1251.850480411" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.593105 4770 scope.go:117] "RemoveContainer" containerID="2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.594566 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.594534252 podStartE2EDuration="2.594534252s" podCreationTimestamp="2025-10-04 03:23:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:00.575833244 +0000 UTC m=+1251.867842966" watchObservedRunningTime="2025-10-04 03:24:00.594534252 +0000 UTC m=+1251.886543964" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.615963 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.627102 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.648749 4770 scope.go:117] "RemoveContainer" containerID="6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e" Oct 04 03:24:00 crc kubenswrapper[4770]: E1004 03:24:00.653316 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e\": container with ID starting with 6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e not found: ID does not exist" containerID="6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.653371 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e"} err="failed to get container status \"6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e\": rpc error: code = NotFound desc = could not find container \"6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e\": container with ID starting with 6d5bc8b507a894a56f26f990e3cdff06efa7315c95eef46eb95dc2eb8d081f8e not found: ID does not exist" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.653404 4770 scope.go:117] "RemoveContainer" containerID="2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624" Oct 04 03:24:00 crc kubenswrapper[4770]: E1004 03:24:00.654110 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624\": container with ID starting with 2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624 not found: ID does not exist" containerID="2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.654162 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624"} err="failed to get container status \"2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624\": rpc error: code = NotFound desc = could not find container \"2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624\": container with ID starting with 2837b12a1b514507285e609666c565461979d0f6dbfc1886187ca77485c44624 not found: ID does not exist" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.654201 4770 scope.go:117] "RemoveContainer" containerID="6633e392a5a6f2b64dfbad7667d1b0267f86fde4b17e7db1b8808041de34c634" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.661139 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.706699 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.717453 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:00 crc kubenswrapper[4770]: E1004 03:24:00.718588 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-log" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.718621 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-log" Oct 04 03:24:00 crc kubenswrapper[4770]: E1004 03:24:00.718636 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a374b7-7611-4f40-b0ae-a6238903e0ea" containerName="kube-state-metrics" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.718645 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a374b7-7611-4f40-b0ae-a6238903e0ea" containerName="kube-state-metrics" Oct 04 03:24:00 crc kubenswrapper[4770]: E1004 03:24:00.718670 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac2690b-33cd-48f3-92d9-b9a0e29442e0" containerName="init" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.718679 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac2690b-33cd-48f3-92d9-b9a0e29442e0" containerName="init" Oct 04 03:24:00 crc kubenswrapper[4770]: E1004 03:24:00.718699 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac2690b-33cd-48f3-92d9-b9a0e29442e0" containerName="dnsmasq-dns" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.718707 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac2690b-33cd-48f3-92d9-b9a0e29442e0" containerName="dnsmasq-dns" Oct 04 03:24:00 crc kubenswrapper[4770]: E1004 03:24:00.718738 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-api" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.718751 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-api" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.719034 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-api" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.719073 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="60a374b7-7611-4f40-b0ae-a6238903e0ea" containerName="kube-state-metrics" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.719094 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ac2690b-33cd-48f3-92d9-b9a0e29442e0" containerName="dnsmasq-dns" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.719106 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" containerName="nova-api-log" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.720913 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.728506 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.733775 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.737026 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.740349 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.741318 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.775381 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.794106 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.887597 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.887714 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvmcg\" (UniqueName: \"kubernetes.io/projected/ddf6a9a5-049a-4656-b425-4ef775a313b9-kube-api-access-wvmcg\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.887751 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf6a9a5-049a-4656-b425-4ef775a313b9-logs\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.887786 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.887808 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.887882 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99r7b\" (UniqueName: \"kubernetes.io/projected/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-api-access-99r7b\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.888187 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.888422 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-config-data\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.990353 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvmcg\" (UniqueName: \"kubernetes.io/projected/ddf6a9a5-049a-4656-b425-4ef775a313b9-kube-api-access-wvmcg\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.990452 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf6a9a5-049a-4656-b425-4ef775a313b9-logs\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.990537 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.990712 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.990819 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99r7b\" (UniqueName: \"kubernetes.io/projected/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-api-access-99r7b\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.990923 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.991292 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf6a9a5-049a-4656-b425-4ef775a313b9-logs\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.991499 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-config-data\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.991621 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.996467 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.996841 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:00 crc kubenswrapper[4770]: I1004 03:24:00.997366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-config-data\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.004787 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.006955 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvmcg\" (UniqueName: \"kubernetes.io/projected/ddf6a9a5-049a-4656-b425-4ef775a313b9-kube-api-access-wvmcg\") pod \"nova-api-0\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " pod="openstack/nova-api-0" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.010760 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.011728 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99r7b\" (UniqueName: \"kubernetes.io/projected/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-api-access-99r7b\") pod \"kube-state-metrics-0\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " pod="openstack/kube-state-metrics-0" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.042054 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.077034 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.593477 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:01 crc kubenswrapper[4770]: W1004 03:24:01.595246 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddf6a9a5_049a_4656_b425_4ef775a313b9.slice/crio-5789ed7c469b92fee7445b697fc041c95f9092c1736a91ee059bce954efcac24 WatchSource:0}: Error finding container 5789ed7c469b92fee7445b697fc041c95f9092c1736a91ee059bce954efcac24: Status 404 returned error can't find the container with id 5789ed7c469b92fee7445b697fc041c95f9092c1736a91ee059bce954efcac24 Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.666623 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.691757 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60a374b7-7611-4f40-b0ae-a6238903e0ea" path="/var/lib/kubelet/pods/60a374b7-7611-4f40-b0ae-a6238903e0ea/volumes" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.692476 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70ea433c-10b4-4fbf-a377-9af4cafd9584" path="/var/lib/kubelet/pods/70ea433c-10b4-4fbf-a377-9af4cafd9584/volumes" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.759264 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.761705 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.796071 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.796167 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.818207 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.819316 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="sg-core" containerID="cri-o://01d04621926a93fa567a19a32594e917816c2218b6b97c3b6209f6220ded3c3d" gracePeriod=30 Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.819367 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="ceilometer-notification-agent" containerID="cri-o://94d10694f73f3942e76d314307eef0e1c606808ce5d537b5bc88766b3a8aece8" gracePeriod=30 Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.819336 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="proxy-httpd" containerID="cri-o://3d9b2277d89aba3b4b5d07d07f7efef56ff6e822cdc71037566424b51ef55257" gracePeriod=30 Oct 04 03:24:01 crc kubenswrapper[4770]: I1004 03:24:01.819590 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="ceilometer-central-agent" containerID="cri-o://ef21a15182a9e6690eea104b5b81ffb153a9e79e9b7fdc87b31ad749a9ad7a7e" gracePeriod=30 Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.615235 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf6a9a5-049a-4656-b425-4ef775a313b9","Type":"ContainerStarted","Data":"e1bece80875de90446ba72c3c5f7b1207f808d04754c1c3049d3868997395326"} Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.615614 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf6a9a5-049a-4656-b425-4ef775a313b9","Type":"ContainerStarted","Data":"697a36aa9b26b0733c87212d1f86d623b70909812a0861bb4ba8d53ce06e0a08"} Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.615625 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf6a9a5-049a-4656-b425-4ef775a313b9","Type":"ContainerStarted","Data":"5789ed7c469b92fee7445b697fc041c95f9092c1736a91ee059bce954efcac24"} Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.623026 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ae4b38bb-7bda-45e9-9d87-f79fe72e9008","Type":"ContainerStarted","Data":"136fdbb6c40c5913eff7c84a5bd828978d5112b321c3e6b34d37ea56010e8b07"} Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.623353 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.623384 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ae4b38bb-7bda-45e9-9d87-f79fe72e9008","Type":"ContainerStarted","Data":"0d0dfd421a79a3b41e1ee5746d2e20b78b0aec53f57e39060355d20191138138"} Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.639422 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.639406689 podStartE2EDuration="2.639406689s" podCreationTimestamp="2025-10-04 03:24:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:02.637331936 +0000 UTC m=+1253.929341658" watchObservedRunningTime="2025-10-04 03:24:02.639406689 +0000 UTC m=+1253.931416401" Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.647834 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerID="3d9b2277d89aba3b4b5d07d07f7efef56ff6e822cdc71037566424b51ef55257" exitCode=0 Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.647865 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerID="01d04621926a93fa567a19a32594e917816c2218b6b97c3b6209f6220ded3c3d" exitCode=2 Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.647874 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerID="94d10694f73f3942e76d314307eef0e1c606808ce5d537b5bc88766b3a8aece8" exitCode=0 Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.647883 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerID="ef21a15182a9e6690eea104b5b81ffb153a9e79e9b7fdc87b31ad749a9ad7a7e" exitCode=0 Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.648744 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerDied","Data":"3d9b2277d89aba3b4b5d07d07f7efef56ff6e822cdc71037566424b51ef55257"} Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.648773 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerDied","Data":"01d04621926a93fa567a19a32594e917816c2218b6b97c3b6209f6220ded3c3d"} Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.648783 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerDied","Data":"94d10694f73f3942e76d314307eef0e1c606808ce5d537b5bc88766b3a8aece8"} Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.648791 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerDied","Data":"ef21a15182a9e6690eea104b5b81ffb153a9e79e9b7fdc87b31ad749a9ad7a7e"} Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.658094 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.254987545 podStartE2EDuration="2.658080298s" podCreationTimestamp="2025-10-04 03:24:00 +0000 UTC" firstStartedPulling="2025-10-04 03:24:01.680319373 +0000 UTC m=+1252.972329085" lastFinishedPulling="2025-10-04 03:24:02.083412126 +0000 UTC m=+1253.375421838" observedRunningTime="2025-10-04 03:24:02.654263128 +0000 UTC m=+1253.946272840" watchObservedRunningTime="2025-10-04 03:24:02.658080298 +0000 UTC m=+1253.950090010" Oct 04 03:24:02 crc kubenswrapper[4770]: I1004 03:24:02.976049 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.142302 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-combined-ca-bundle\") pod \"f4edadec-fa14-45b6-9a4f-83c328b8e962\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.142392 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-scripts\") pod \"f4edadec-fa14-45b6-9a4f-83c328b8e962\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.142461 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-log-httpd\") pod \"f4edadec-fa14-45b6-9a4f-83c328b8e962\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.142507 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-sg-core-conf-yaml\") pod \"f4edadec-fa14-45b6-9a4f-83c328b8e962\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.142529 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-config-data\") pod \"f4edadec-fa14-45b6-9a4f-83c328b8e962\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.142664 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk9jh\" (UniqueName: \"kubernetes.io/projected/f4edadec-fa14-45b6-9a4f-83c328b8e962-kube-api-access-qk9jh\") pod \"f4edadec-fa14-45b6-9a4f-83c328b8e962\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.142702 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-run-httpd\") pod \"f4edadec-fa14-45b6-9a4f-83c328b8e962\" (UID: \"f4edadec-fa14-45b6-9a4f-83c328b8e962\") " Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.143217 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f4edadec-fa14-45b6-9a4f-83c328b8e962" (UID: "f4edadec-fa14-45b6-9a4f-83c328b8e962"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.143421 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f4edadec-fa14-45b6-9a4f-83c328b8e962" (UID: "f4edadec-fa14-45b6-9a4f-83c328b8e962"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.143761 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.143789 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f4edadec-fa14-45b6-9a4f-83c328b8e962-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.149212 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-scripts" (OuterVolumeSpecName: "scripts") pod "f4edadec-fa14-45b6-9a4f-83c328b8e962" (UID: "f4edadec-fa14-45b6-9a4f-83c328b8e962"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.149347 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4edadec-fa14-45b6-9a4f-83c328b8e962-kube-api-access-qk9jh" (OuterVolumeSpecName: "kube-api-access-qk9jh") pod "f4edadec-fa14-45b6-9a4f-83c328b8e962" (UID: "f4edadec-fa14-45b6-9a4f-83c328b8e962"). InnerVolumeSpecName "kube-api-access-qk9jh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.184993 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f4edadec-fa14-45b6-9a4f-83c328b8e962" (UID: "f4edadec-fa14-45b6-9a4f-83c328b8e962"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.230635 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4edadec-fa14-45b6-9a4f-83c328b8e962" (UID: "f4edadec-fa14-45b6-9a4f-83c328b8e962"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.246139 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk9jh\" (UniqueName: \"kubernetes.io/projected/f4edadec-fa14-45b6-9a4f-83c328b8e962-kube-api-access-qk9jh\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.246184 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.246196 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.246209 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.256853 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-config-data" (OuterVolumeSpecName: "config-data") pod "f4edadec-fa14-45b6-9a4f-83c328b8e962" (UID: "f4edadec-fa14-45b6-9a4f-83c328b8e962"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.348284 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4edadec-fa14-45b6-9a4f-83c328b8e962-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.671933 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f4edadec-fa14-45b6-9a4f-83c328b8e962","Type":"ContainerDied","Data":"1302a41daa208af21523723ba6416b9d9372becea8f9071fbe29343ef2de2878"} Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.672270 4770 scope.go:117] "RemoveContainer" containerID="3d9b2277d89aba3b4b5d07d07f7efef56ff6e822cdc71037566424b51ef55257" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.674535 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.699307 4770 scope.go:117] "RemoveContainer" containerID="01d04621926a93fa567a19a32594e917816c2218b6b97c3b6209f6220ded3c3d" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.725538 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.742786 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.750707 4770 scope.go:117] "RemoveContainer" containerID="94d10694f73f3942e76d314307eef0e1c606808ce5d537b5bc88766b3a8aece8" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.770536 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:03 crc kubenswrapper[4770]: E1004 03:24:03.771092 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="ceilometer-notification-agent" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.771116 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="ceilometer-notification-agent" Oct 04 03:24:03 crc kubenswrapper[4770]: E1004 03:24:03.771139 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="proxy-httpd" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.771147 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="proxy-httpd" Oct 04 03:24:03 crc kubenswrapper[4770]: E1004 03:24:03.771174 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="ceilometer-central-agent" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.771181 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="ceilometer-central-agent" Oct 04 03:24:03 crc kubenswrapper[4770]: E1004 03:24:03.771212 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="sg-core" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.771221 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="sg-core" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.771515 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="proxy-httpd" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.771538 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="sg-core" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.771547 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="ceilometer-central-agent" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.771557 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" containerName="ceilometer-notification-agent" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.774332 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.776570 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.776720 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.777030 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.780254 4770 scope.go:117] "RemoveContainer" containerID="ef21a15182a9e6690eea104b5b81ffb153a9e79e9b7fdc87b31ad749a9ad7a7e" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.781068 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.869685 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.869783 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plj6q\" (UniqueName: \"kubernetes.io/projected/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-kube-api-access-plj6q\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.869813 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-run-httpd\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.869919 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-scripts\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.869994 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-config-data\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.870143 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-log-httpd\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.870201 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.870378 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.957655 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.972376 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-run-httpd\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.972832 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plj6q\" (UniqueName: \"kubernetes.io/projected/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-kube-api-access-plj6q\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.972788 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-run-httpd\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.972913 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-scripts\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.972949 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-config-data\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.972996 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-log-httpd\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.973655 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.973495 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-log-httpd\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.973795 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.973891 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.979333 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.979595 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.980271 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-config-data\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.980749 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:03 crc kubenswrapper[4770]: I1004 03:24:03.987721 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-scripts\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:04 crc kubenswrapper[4770]: I1004 03:24:04.010842 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plj6q\" (UniqueName: \"kubernetes.io/projected/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-kube-api-access-plj6q\") pod \"ceilometer-0\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " pod="openstack/ceilometer-0" Oct 04 03:24:04 crc kubenswrapper[4770]: I1004 03:24:04.104264 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:24:04 crc kubenswrapper[4770]: W1004 03:24:04.607323 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a2cb5a0_b4a3_4d14_ad1a_175ad1d80940.slice/crio-186262b06ced0802794e3a0db1db6ada06b93d46e55c3c07841aa10a4149ede1 WatchSource:0}: Error finding container 186262b06ced0802794e3a0db1db6ada06b93d46e55c3c07841aa10a4149ede1: Status 404 returned error can't find the container with id 186262b06ced0802794e3a0db1db6ada06b93d46e55c3c07841aa10a4149ede1 Oct 04 03:24:04 crc kubenswrapper[4770]: I1004 03:24:04.609906 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:04 crc kubenswrapper[4770]: I1004 03:24:04.661475 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="60a374b7-7611-4f40-b0ae-a6238903e0ea" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:04 crc kubenswrapper[4770]: I1004 03:24:04.686541 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerStarted","Data":"186262b06ced0802794e3a0db1db6ada06b93d46e55c3c07841aa10a4149ede1"} Oct 04 03:24:05 crc kubenswrapper[4770]: I1004 03:24:05.685875 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4edadec-fa14-45b6-9a4f-83c328b8e962" path="/var/lib/kubelet/pods/f4edadec-fa14-45b6-9a4f-83c328b8e962/volumes" Oct 04 03:24:05 crc kubenswrapper[4770]: I1004 03:24:05.701584 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerStarted","Data":"9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12"} Oct 04 03:24:06 crc kubenswrapper[4770]: I1004 03:24:06.715425 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerStarted","Data":"568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d"} Oct 04 03:24:06 crc kubenswrapper[4770]: I1004 03:24:06.758625 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 03:24:06 crc kubenswrapper[4770]: I1004 03:24:06.758693 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 03:24:07 crc kubenswrapper[4770]: I1004 03:24:07.738541 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerStarted","Data":"3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c"} Oct 04 03:24:07 crc kubenswrapper[4770]: I1004 03:24:07.774200 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:07 crc kubenswrapper[4770]: I1004 03:24:07.774224 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:08 crc kubenswrapper[4770]: I1004 03:24:08.753216 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerStarted","Data":"b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43"} Oct 04 03:24:08 crc kubenswrapper[4770]: I1004 03:24:08.756376 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 03:24:08 crc kubenswrapper[4770]: I1004 03:24:08.782617 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.077988386 podStartE2EDuration="5.782600277s" podCreationTimestamp="2025-10-04 03:24:03 +0000 UTC" firstStartedPulling="2025-10-04 03:24:04.611079913 +0000 UTC m=+1255.903089625" lastFinishedPulling="2025-10-04 03:24:08.315691804 +0000 UTC m=+1259.607701516" observedRunningTime="2025-10-04 03:24:08.781386185 +0000 UTC m=+1260.073395897" watchObservedRunningTime="2025-10-04 03:24:08.782600277 +0000 UTC m=+1260.074609989" Oct 04 03:24:08 crc kubenswrapper[4770]: I1004 03:24:08.873854 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 03:24:08 crc kubenswrapper[4770]: I1004 03:24:08.958025 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 03:24:09 crc kubenswrapper[4770]: I1004 03:24:09.011215 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 03:24:09 crc kubenswrapper[4770]: I1004 03:24:09.790816 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 03:24:11 crc kubenswrapper[4770]: I1004 03:24:11.043376 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 03:24:11 crc kubenswrapper[4770]: I1004 03:24:11.043880 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 03:24:11 crc kubenswrapper[4770]: I1004 03:24:11.096166 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 03:24:12 crc kubenswrapper[4770]: I1004 03:24:12.125339 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:12 crc kubenswrapper[4770]: I1004 03:24:12.125367 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:16 crc kubenswrapper[4770]: I1004 03:24:16.764288 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 03:24:16 crc kubenswrapper[4770]: I1004 03:24:16.766167 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 03:24:16 crc kubenswrapper[4770]: I1004 03:24:16.770508 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 03:24:16 crc kubenswrapper[4770]: I1004 03:24:16.855922 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.633394 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.743762 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpfxf\" (UniqueName: \"kubernetes.io/projected/e79627e4-ecd2-48f3-90c2-64c68bf495f9-kube-api-access-wpfxf\") pod \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.743892 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-config-data\") pod \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.743946 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-combined-ca-bundle\") pod \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\" (UID: \"e79627e4-ecd2-48f3-90c2-64c68bf495f9\") " Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.752934 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e79627e4-ecd2-48f3-90c2-64c68bf495f9-kube-api-access-wpfxf" (OuterVolumeSpecName: "kube-api-access-wpfxf") pod "e79627e4-ecd2-48f3-90c2-64c68bf495f9" (UID: "e79627e4-ecd2-48f3-90c2-64c68bf495f9"). InnerVolumeSpecName "kube-api-access-wpfxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.779499 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-config-data" (OuterVolumeSpecName: "config-data") pod "e79627e4-ecd2-48f3-90c2-64c68bf495f9" (UID: "e79627e4-ecd2-48f3-90c2-64c68bf495f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.802442 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e79627e4-ecd2-48f3-90c2-64c68bf495f9" (UID: "e79627e4-ecd2-48f3-90c2-64c68bf495f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.846406 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpfxf\" (UniqueName: \"kubernetes.io/projected/e79627e4-ecd2-48f3-90c2-64c68bf495f9-kube-api-access-wpfxf\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.846440 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.846450 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79627e4-ecd2-48f3-90c2-64c68bf495f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.893544 4770 generic.go:334] "Generic (PLEG): container finished" podID="e79627e4-ecd2-48f3-90c2-64c68bf495f9" containerID="de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d" exitCode=137 Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.893653 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.893650 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e79627e4-ecd2-48f3-90c2-64c68bf495f9","Type":"ContainerDied","Data":"de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d"} Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.894119 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e79627e4-ecd2-48f3-90c2-64c68bf495f9","Type":"ContainerDied","Data":"679280c5e7ced30a83dd00bc0cd88ddd106f6d9b6be82969897f7d878c17d052"} Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.894176 4770 scope.go:117] "RemoveContainer" containerID="de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.920670 4770 scope.go:117] "RemoveContainer" containerID="de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d" Oct 04 03:24:20 crc kubenswrapper[4770]: E1004 03:24:20.921364 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d\": container with ID starting with de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d not found: ID does not exist" containerID="de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.921508 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d"} err="failed to get container status \"de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d\": rpc error: code = NotFound desc = could not find container \"de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d\": container with ID starting with de18cf85f99311f80bdd114cd24b083b76cbd23eea9e3d73c14d286376793f4d not found: ID does not exist" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.955695 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.976060 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.989830 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:24:20 crc kubenswrapper[4770]: E1004 03:24:20.990462 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e79627e4-ecd2-48f3-90c2-64c68bf495f9" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.990490 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e79627e4-ecd2-48f3-90c2-64c68bf495f9" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.990968 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e79627e4-ecd2-48f3-90c2-64c68bf495f9" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.992099 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.995624 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.995834 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 03:24:20 crc kubenswrapper[4770]: I1004 03:24:20.995835 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.000558 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.048296 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.048371 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.049013 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.049033 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.051949 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.054039 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.152055 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.152105 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phl9r\" (UniqueName: \"kubernetes.io/projected/1d236ba3-55db-460a-9daa-d6e6e8b06766-kube-api-access-phl9r\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.152417 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.152496 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.152856 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.228172 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-x9khq"] Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.230327 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.239494 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-x9khq"] Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.254931 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.254978 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.255052 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.255132 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.255158 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phl9r\" (UniqueName: \"kubernetes.io/projected/1d236ba3-55db-460a-9daa-d6e6e8b06766-kube-api-access-phl9r\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.262036 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.273269 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.277605 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phl9r\" (UniqueName: \"kubernetes.io/projected/1d236ba3-55db-460a-9daa-d6e6e8b06766-kube-api-access-phl9r\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.283925 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.284155 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.313869 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.357035 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.357113 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4dpx\" (UniqueName: \"kubernetes.io/projected/4c47d914-960a-4c5e-a6f3-584998910e09-kube-api-access-p4dpx\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.357483 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-svc\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.357670 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-swift-storage-0\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.357726 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-config\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.357818 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.461761 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.462217 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4dpx\" (UniqueName: \"kubernetes.io/projected/4c47d914-960a-4c5e-a6f3-584998910e09-kube-api-access-p4dpx\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.462312 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-svc\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.462385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-swift-storage-0\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.462420 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-config\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.462467 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.463925 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-swift-storage-0\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.464485 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-svc\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.465023 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-config\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.466111 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.466633 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.499577 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4dpx\" (UniqueName: \"kubernetes.io/projected/4c47d914-960a-4c5e-a6f3-584998910e09-kube-api-access-p4dpx\") pod \"dnsmasq-dns-6c5b8b845f-x9khq\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.684508 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e79627e4-ecd2-48f3-90c2-64c68bf495f9" path="/var/lib/kubelet/pods/e79627e4-ecd2-48f3-90c2-64c68bf495f9/volumes" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.721874 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.838484 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:24:21 crc kubenswrapper[4770]: I1004 03:24:21.921180 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1d236ba3-55db-460a-9daa-d6e6e8b06766","Type":"ContainerStarted","Data":"4b1b87141144b8d9ae970d74e97c3a2a03a3f7e4342814bc56ce4a8bfb6fc71c"} Oct 04 03:24:22 crc kubenswrapper[4770]: I1004 03:24:22.214297 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-x9khq"] Oct 04 03:24:22 crc kubenswrapper[4770]: I1004 03:24:22.930495 4770 generic.go:334] "Generic (PLEG): container finished" podID="4c47d914-960a-4c5e-a6f3-584998910e09" containerID="2a8d8f9a02c3dd2f0f0adf8b24176660cc839972d5330bc42bef1b971ce6c49b" exitCode=0 Oct 04 03:24:22 crc kubenswrapper[4770]: I1004 03:24:22.930588 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" event={"ID":"4c47d914-960a-4c5e-a6f3-584998910e09","Type":"ContainerDied","Data":"2a8d8f9a02c3dd2f0f0adf8b24176660cc839972d5330bc42bef1b971ce6c49b"} Oct 04 03:24:22 crc kubenswrapper[4770]: I1004 03:24:22.930908 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" event={"ID":"4c47d914-960a-4c5e-a6f3-584998910e09","Type":"ContainerStarted","Data":"0a9006b887e7f4386afae8b76b8ef40d5a366866c54ce68deaebb74622b739fc"} Oct 04 03:24:22 crc kubenswrapper[4770]: I1004 03:24:22.932336 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1d236ba3-55db-460a-9daa-d6e6e8b06766","Type":"ContainerStarted","Data":"f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8"} Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.025162 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.025136157 podStartE2EDuration="3.025136157s" podCreationTimestamp="2025-10-04 03:24:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:22.997278078 +0000 UTC m=+1274.289287790" watchObservedRunningTime="2025-10-04 03:24:23.025136157 +0000 UTC m=+1274.317145869" Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.695240 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.696456 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="ceilometer-central-agent" containerID="cri-o://9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12" gracePeriod=30 Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.698653 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="proxy-httpd" containerID="cri-o://b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43" gracePeriod=30 Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.698841 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="sg-core" containerID="cri-o://3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c" gracePeriod=30 Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.698917 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="ceilometer-notification-agent" containerID="cri-o://568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d" gracePeriod=30 Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.709732 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": EOF" Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.945288 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" event={"ID":"4c47d914-960a-4c5e-a6f3-584998910e09","Type":"ContainerStarted","Data":"e953a76ab2110b06b5538727038bb150355effbb23a98d4d39a75531a080b3b0"} Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.946548 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.953654 4770 generic.go:334] "Generic (PLEG): container finished" podID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerID="b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43" exitCode=0 Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.953695 4770 generic.go:334] "Generic (PLEG): container finished" podID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerID="3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c" exitCode=2 Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.953735 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerDied","Data":"b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43"} Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.953808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerDied","Data":"3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c"} Oct 04 03:24:23 crc kubenswrapper[4770]: I1004 03:24:23.985150 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" podStartSLOduration=2.985130948 podStartE2EDuration="2.985130948s" podCreationTimestamp="2025-10-04 03:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:23.975134426 +0000 UTC m=+1275.267144138" watchObservedRunningTime="2025-10-04 03:24:23.985130948 +0000 UTC m=+1275.277140660" Oct 04 03:24:24 crc kubenswrapper[4770]: I1004 03:24:24.552530 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:24 crc kubenswrapper[4770]: I1004 03:24:24.553131 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-log" containerID="cri-o://697a36aa9b26b0733c87212d1f86d623b70909812a0861bb4ba8d53ce06e0a08" gracePeriod=30 Oct 04 03:24:24 crc kubenswrapper[4770]: I1004 03:24:24.553181 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-api" containerID="cri-o://e1bece80875de90446ba72c3c5f7b1207f808d04754c1c3049d3868997395326" gracePeriod=30 Oct 04 03:24:24 crc kubenswrapper[4770]: I1004 03:24:24.964996 4770 generic.go:334] "Generic (PLEG): container finished" podID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerID="9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12" exitCode=0 Oct 04 03:24:24 crc kubenswrapper[4770]: I1004 03:24:24.965087 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerDied","Data":"9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12"} Oct 04 03:24:24 crc kubenswrapper[4770]: I1004 03:24:24.967383 4770 generic.go:334] "Generic (PLEG): container finished" podID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerID="697a36aa9b26b0733c87212d1f86d623b70909812a0861bb4ba8d53ce06e0a08" exitCode=143 Oct 04 03:24:24 crc kubenswrapper[4770]: I1004 03:24:24.967455 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf6a9a5-049a-4656-b425-4ef775a313b9","Type":"ContainerDied","Data":"697a36aa9b26b0733c87212d1f86d623b70909812a0861bb4ba8d53ce06e0a08"} Oct 04 03:24:26 crc kubenswrapper[4770]: I1004 03:24:26.314631 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.629249 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.697254 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-ceilometer-tls-certs\") pod \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.697364 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-log-httpd\") pod \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.697452 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plj6q\" (UniqueName: \"kubernetes.io/projected/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-kube-api-access-plj6q\") pod \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.697483 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-config-data\") pod \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.697508 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-sg-core-conf-yaml\") pod \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.697568 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-run-httpd\") pod \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.697623 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-combined-ca-bundle\") pod \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.697649 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-scripts\") pod \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\" (UID: \"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940\") " Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.698585 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" (UID: "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.701641 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" (UID: "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.707310 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-scripts" (OuterVolumeSpecName: "scripts") pod "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" (UID: "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.710822 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-kube-api-access-plj6q" (OuterVolumeSpecName: "kube-api-access-plj6q") pod "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" (UID: "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940"). InnerVolumeSpecName "kube-api-access-plj6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.743235 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" (UID: "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.775711 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" (UID: "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.800483 4770 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.800649 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.800722 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plj6q\" (UniqueName: \"kubernetes.io/projected/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-kube-api-access-plj6q\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.800792 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.800859 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.800918 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.817749 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" (UID: "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.838271 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-config-data" (OuterVolumeSpecName: "config-data") pod "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" (UID: "1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.904612 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:27 crc kubenswrapper[4770]: I1004 03:24:27.904658 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.003054 4770 generic.go:334] "Generic (PLEG): container finished" podID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerID="e1bece80875de90446ba72c3c5f7b1207f808d04754c1c3049d3868997395326" exitCode=0 Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.003112 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf6a9a5-049a-4656-b425-4ef775a313b9","Type":"ContainerDied","Data":"e1bece80875de90446ba72c3c5f7b1207f808d04754c1c3049d3868997395326"} Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.008920 4770 generic.go:334] "Generic (PLEG): container finished" podID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerID="568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d" exitCode=0 Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.008977 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerDied","Data":"568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d"} Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.009022 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940","Type":"ContainerDied","Data":"186262b06ced0802794e3a0db1db6ada06b93d46e55c3c07841aa10a4149ede1"} Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.009046 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.009078 4770 scope.go:117] "RemoveContainer" containerID="b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.052492 4770 scope.go:117] "RemoveContainer" containerID="3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.064857 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.080080 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.105840 4770 scope.go:117] "RemoveContainer" containerID="568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.119065 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:28 crc kubenswrapper[4770]: E1004 03:24:28.119714 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="ceilometer-notification-agent" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.119760 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="ceilometer-notification-agent" Oct 04 03:24:28 crc kubenswrapper[4770]: E1004 03:24:28.119780 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="ceilometer-central-agent" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.119795 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="ceilometer-central-agent" Oct 04 03:24:28 crc kubenswrapper[4770]: E1004 03:24:28.119813 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="proxy-httpd" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.119825 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="proxy-httpd" Oct 04 03:24:28 crc kubenswrapper[4770]: E1004 03:24:28.119891 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="sg-core" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.119903 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="sg-core" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.120253 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="sg-core" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.120293 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="proxy-httpd" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.120311 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="ceilometer-notification-agent" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.120341 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" containerName="ceilometer-central-agent" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.123834 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.125810 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.131274 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.131360 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.138799 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.162792 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.171863 4770 scope.go:117] "RemoveContainer" containerID="9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.215934 4770 scope.go:117] "RemoveContainer" containerID="b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43" Oct 04 03:24:28 crc kubenswrapper[4770]: E1004 03:24:28.217989 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43\": container with ID starting with b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43 not found: ID does not exist" containerID="b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.218094 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43"} err="failed to get container status \"b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43\": rpc error: code = NotFound desc = could not find container \"b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43\": container with ID starting with b786c36f77cfa1de1eae0cf30ccd059609632c2f14875c50a054f2057dca5b43 not found: ID does not exist" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.218122 4770 scope.go:117] "RemoveContainer" containerID="3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c" Oct 04 03:24:28 crc kubenswrapper[4770]: E1004 03:24:28.219114 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c\": container with ID starting with 3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c not found: ID does not exist" containerID="3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.219169 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c"} err="failed to get container status \"3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c\": rpc error: code = NotFound desc = could not find container \"3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c\": container with ID starting with 3e8ad62c355121a882f5e24f93ec21995cd7dd821dfd73a356a847016a1b875c not found: ID does not exist" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.219242 4770 scope.go:117] "RemoveContainer" containerID="568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d" Oct 04 03:24:28 crc kubenswrapper[4770]: E1004 03:24:28.220341 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d\": container with ID starting with 568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d not found: ID does not exist" containerID="568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.220395 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d"} err="failed to get container status \"568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d\": rpc error: code = NotFound desc = could not find container \"568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d\": container with ID starting with 568824c9810a85035827eda428ae4e861a9df434ad18eb26f1d069ae88425c2d not found: ID does not exist" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.220411 4770 scope.go:117] "RemoveContainer" containerID="9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12" Oct 04 03:24:28 crc kubenswrapper[4770]: E1004 03:24:28.220613 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12\": container with ID starting with 9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12 not found: ID does not exist" containerID="9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.220635 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12"} err="failed to get container status \"9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12\": rpc error: code = NotFound desc = could not find container \"9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12\": container with ID starting with 9856967edf8dcbcbb3754caf5b4b11f231b3bb6e934ce5bd3784da07f8099e12 not found: ID does not exist" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.226500 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-log-httpd\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.226569 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-config-data\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.226674 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-scripts\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.226727 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.226881 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.227056 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.227098 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-run-httpd\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.227166 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzp9q\" (UniqueName: \"kubernetes.io/projected/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-kube-api-access-fzp9q\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.329265 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-config-data\") pod \"ddf6a9a5-049a-4656-b425-4ef775a313b9\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.329558 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf6a9a5-049a-4656-b425-4ef775a313b9-logs\") pod \"ddf6a9a5-049a-4656-b425-4ef775a313b9\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.329594 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-combined-ca-bundle\") pod \"ddf6a9a5-049a-4656-b425-4ef775a313b9\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.329663 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvmcg\" (UniqueName: \"kubernetes.io/projected/ddf6a9a5-049a-4656-b425-4ef775a313b9-kube-api-access-wvmcg\") pod \"ddf6a9a5-049a-4656-b425-4ef775a313b9\" (UID: \"ddf6a9a5-049a-4656-b425-4ef775a313b9\") " Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.332396 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.332467 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-run-httpd\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.332547 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzp9q\" (UniqueName: \"kubernetes.io/projected/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-kube-api-access-fzp9q\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.332794 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-log-httpd\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.332878 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-config-data\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.332944 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-scripts\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.333000 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.333139 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.334655 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-run-httpd\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.337201 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-log-httpd\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.337779 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddf6a9a5-049a-4656-b425-4ef775a313b9-logs" (OuterVolumeSpecName: "logs") pod "ddf6a9a5-049a-4656-b425-4ef775a313b9" (UID: "ddf6a9a5-049a-4656-b425-4ef775a313b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.338715 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-config-data\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.345578 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.348243 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddf6a9a5-049a-4656-b425-4ef775a313b9-kube-api-access-wvmcg" (OuterVolumeSpecName: "kube-api-access-wvmcg") pod "ddf6a9a5-049a-4656-b425-4ef775a313b9" (UID: "ddf6a9a5-049a-4656-b425-4ef775a313b9"). InnerVolumeSpecName "kube-api-access-wvmcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.352781 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.353291 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-scripts\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.355866 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzp9q\" (UniqueName: \"kubernetes.io/projected/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-kube-api-access-fzp9q\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.370158 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.385105 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddf6a9a5-049a-4656-b425-4ef775a313b9" (UID: "ddf6a9a5-049a-4656-b425-4ef775a313b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.402387 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-config-data" (OuterVolumeSpecName: "config-data") pod "ddf6a9a5-049a-4656-b425-4ef775a313b9" (UID: "ddf6a9a5-049a-4656-b425-4ef775a313b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.435255 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.435316 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf6a9a5-049a-4656-b425-4ef775a313b9-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.435331 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf6a9a5-049a-4656-b425-4ef775a313b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.435354 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvmcg\" (UniqueName: \"kubernetes.io/projected/ddf6a9a5-049a-4656-b425-4ef775a313b9-kube-api-access-wvmcg\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.446215 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.960270 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:24:28 crc kubenswrapper[4770]: I1004 03:24:28.968126 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.022426 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerStarted","Data":"cbe9489fa8c679d8b66c4b01c781525a45f303d607c8209d7b432b97f8583077"} Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.025638 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf6a9a5-049a-4656-b425-4ef775a313b9","Type":"ContainerDied","Data":"5789ed7c469b92fee7445b697fc041c95f9092c1736a91ee059bce954efcac24"} Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.025673 4770 scope.go:117] "RemoveContainer" containerID="e1bece80875de90446ba72c3c5f7b1207f808d04754c1c3049d3868997395326" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.025768 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.054404 4770 scope.go:117] "RemoveContainer" containerID="697a36aa9b26b0733c87212d1f86d623b70909812a0861bb4ba8d53ce06e0a08" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.069021 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.082098 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.105886 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:29 crc kubenswrapper[4770]: E1004 03:24:29.106805 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-api" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.106844 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-api" Oct 04 03:24:29 crc kubenswrapper[4770]: E1004 03:24:29.106888 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-log" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.106901 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-log" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.107283 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-api" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.107318 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" containerName="nova-api-log" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.109279 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.111076 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.112049 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.113259 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.137432 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.253953 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.254102 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.254161 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd93c46b-fdd0-44ac-a503-673d32a72e27-logs\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.254193 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-public-tls-certs\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.254346 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-config-data\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.254375 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbfnr\" (UniqueName: \"kubernetes.io/projected/dd93c46b-fdd0-44ac-a503-673d32a72e27-kube-api-access-gbfnr\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.357527 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.357815 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.357922 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd93c46b-fdd0-44ac-a503-673d32a72e27-logs\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.357949 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-public-tls-certs\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.358333 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-config-data\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.358372 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbfnr\" (UniqueName: \"kubernetes.io/projected/dd93c46b-fdd0-44ac-a503-673d32a72e27-kube-api-access-gbfnr\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.358409 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd93c46b-fdd0-44ac-a503-673d32a72e27-logs\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.364588 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-public-tls-certs\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.365680 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.365913 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.367725 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-config-data\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.382989 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbfnr\" (UniqueName: \"kubernetes.io/projected/dd93c46b-fdd0-44ac-a503-673d32a72e27-kube-api-access-gbfnr\") pod \"nova-api-0\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.429511 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.704422 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940" path="/var/lib/kubelet/pods/1a2cb5a0-b4a3-4d14-ad1a-175ad1d80940/volumes" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.707761 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddf6a9a5-049a-4656-b425-4ef775a313b9" path="/var/lib/kubelet/pods/ddf6a9a5-049a-4656-b425-4ef775a313b9/volumes" Oct 04 03:24:29 crc kubenswrapper[4770]: I1004 03:24:29.952193 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:29 crc kubenswrapper[4770]: W1004 03:24:29.957241 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd93c46b_fdd0_44ac_a503_673d32a72e27.slice/crio-f824d739053c047ad163b5150e861626f83dfb1106f93593c99452b944bed303 WatchSource:0}: Error finding container f824d739053c047ad163b5150e861626f83dfb1106f93593c99452b944bed303: Status 404 returned error can't find the container with id f824d739053c047ad163b5150e861626f83dfb1106f93593c99452b944bed303 Oct 04 03:24:30 crc kubenswrapper[4770]: I1004 03:24:30.045449 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerStarted","Data":"3eef293d1922aeebed6a824d6768b260279f3c458835051702da305ac78ec1e4"} Oct 04 03:24:30 crc kubenswrapper[4770]: I1004 03:24:30.047116 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd93c46b-fdd0-44ac-a503-673d32a72e27","Type":"ContainerStarted","Data":"f824d739053c047ad163b5150e861626f83dfb1106f93593c99452b944bed303"} Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.064836 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerStarted","Data":"6fd96f4fbeb0043266b189a92e71e234f2a99b6b94cea0a2db77c70f4d686a5b"} Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.067654 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd93c46b-fdd0-44ac-a503-673d32a72e27","Type":"ContainerStarted","Data":"9fc4a0761d611105cc75c38fdc3d7d9340739f44d9017cc1b89853a831f1ee49"} Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.067711 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd93c46b-fdd0-44ac-a503-673d32a72e27","Type":"ContainerStarted","Data":"4df76b0965987281d66940d30bfcbf3da14ec8a918a94e04279e03639ad4bdaf"} Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.107850 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.107824615 podStartE2EDuration="2.107824615s" podCreationTimestamp="2025-10-04 03:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:31.091780376 +0000 UTC m=+1282.383790108" watchObservedRunningTime="2025-10-04 03:24:31.107824615 +0000 UTC m=+1282.399834357" Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.314531 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.342036 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.723214 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.802654 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.802722 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.802782 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.803358 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6b0b7a6fef39110ff7706ef744d6f617ed372a11387e2977774d987a4194d8f"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.803441 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://d6b0b7a6fef39110ff7706ef744d6f617ed372a11387e2977774d987a4194d8f" gracePeriod=600 Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.833689 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-47pc5"] Oct 04 03:24:31 crc kubenswrapper[4770]: I1004 03:24:31.834089 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" podUID="11acd15b-b791-4869-a16b-37d6030b2079" containerName="dnsmasq-dns" containerID="cri-o://8202959a2d6a9e3319d4f040fed6d702380bf5fbfc420149e5eda797c7472b95" gracePeriod=10 Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.105136 4770 generic.go:334] "Generic (PLEG): container finished" podID="11acd15b-b791-4869-a16b-37d6030b2079" containerID="8202959a2d6a9e3319d4f040fed6d702380bf5fbfc420149e5eda797c7472b95" exitCode=0 Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.105491 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" event={"ID":"11acd15b-b791-4869-a16b-37d6030b2079","Type":"ContainerDied","Data":"8202959a2d6a9e3319d4f040fed6d702380bf5fbfc420149e5eda797c7472b95"} Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.112041 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="d6b0b7a6fef39110ff7706ef744d6f617ed372a11387e2977774d987a4194d8f" exitCode=0 Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.112108 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"d6b0b7a6fef39110ff7706ef744d6f617ed372a11387e2977774d987a4194d8f"} Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.112143 4770 scope.go:117] "RemoveContainer" containerID="73d28cffdd4e71105ae4d9aa35d59b331f61e307981b18b94799778a5b0a5a8d" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.123539 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerStarted","Data":"3ebcae43ad55173cedbd5766bb36e217f05f019084426c4bf34462bc4520a90f"} Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.152563 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:24:32 crc kubenswrapper[4770]: E1004 03:24:32.289985 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bd3be93_1791_4cd1_b3ae_b4032548e93a.slice/crio-d6b0b7a6fef39110ff7706ef744d6f617ed372a11387e2977774d987a4194d8f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11acd15b_b791_4869_a16b_37d6030b2079.slice/crio-conmon-8202959a2d6a9e3319d4f040fed6d702380bf5fbfc420149e5eda797c7472b95.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11acd15b_b791_4869_a16b_37d6030b2079.slice/crio-8202959a2d6a9e3319d4f040fed6d702380bf5fbfc420149e5eda797c7472b95.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bd3be93_1791_4cd1_b3ae_b4032548e93a.slice/crio-conmon-d6b0b7a6fef39110ff7706ef744d6f617ed372a11387e2977774d987a4194d8f.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.368616 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-4rvtr"] Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.372235 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.377412 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.377762 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.387384 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4rvtr"] Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.500867 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.529466 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.529867 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvg4f\" (UniqueName: \"kubernetes.io/projected/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-kube-api-access-gvg4f\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.529908 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-scripts\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.529959 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-config-data\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.631625 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-swift-storage-0\") pod \"11acd15b-b791-4869-a16b-37d6030b2079\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.631791 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-sb\") pod \"11acd15b-b791-4869-a16b-37d6030b2079\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.631836 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq8sg\" (UniqueName: \"kubernetes.io/projected/11acd15b-b791-4869-a16b-37d6030b2079-kube-api-access-lq8sg\") pod \"11acd15b-b791-4869-a16b-37d6030b2079\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.631876 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-nb\") pod \"11acd15b-b791-4869-a16b-37d6030b2079\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.631917 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-svc\") pod \"11acd15b-b791-4869-a16b-37d6030b2079\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.631957 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-config\") pod \"11acd15b-b791-4869-a16b-37d6030b2079\" (UID: \"11acd15b-b791-4869-a16b-37d6030b2079\") " Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.632391 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-config-data\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.632515 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.632593 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvg4f\" (UniqueName: \"kubernetes.io/projected/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-kube-api-access-gvg4f\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.632640 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-scripts\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.639915 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-config-data\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.641133 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-scripts\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.659869 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.661215 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11acd15b-b791-4869-a16b-37d6030b2079-kube-api-access-lq8sg" (OuterVolumeSpecName: "kube-api-access-lq8sg") pod "11acd15b-b791-4869-a16b-37d6030b2079" (UID: "11acd15b-b791-4869-a16b-37d6030b2079"). InnerVolumeSpecName "kube-api-access-lq8sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.661463 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvg4f\" (UniqueName: \"kubernetes.io/projected/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-kube-api-access-gvg4f\") pod \"nova-cell1-cell-mapping-4rvtr\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.707718 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "11acd15b-b791-4869-a16b-37d6030b2079" (UID: "11acd15b-b791-4869-a16b-37d6030b2079"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.714671 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "11acd15b-b791-4869-a16b-37d6030b2079" (UID: "11acd15b-b791-4869-a16b-37d6030b2079"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.729604 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11acd15b-b791-4869-a16b-37d6030b2079" (UID: "11acd15b-b791-4869-a16b-37d6030b2079"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.734786 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.734824 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.734835 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq8sg\" (UniqueName: \"kubernetes.io/projected/11acd15b-b791-4869-a16b-37d6030b2079-kube-api-access-lq8sg\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.734848 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.735500 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-config" (OuterVolumeSpecName: "config") pod "11acd15b-b791-4869-a16b-37d6030b2079" (UID: "11acd15b-b791-4869-a16b-37d6030b2079"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.740781 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11acd15b-b791-4869-a16b-37d6030b2079" (UID: "11acd15b-b791-4869-a16b-37d6030b2079"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.799234 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.837657 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:32 crc kubenswrapper[4770]: I1004 03:24:32.838130 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11acd15b-b791-4869-a16b-37d6030b2079-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.137064 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerStarted","Data":"f6cc4fe8efa2daaa6113859b79bc85ffd8f259a17e76cdbe64107c051e5c6660"} Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.140204 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.147734 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.147757 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6d46f6cf-47pc5" event={"ID":"11acd15b-b791-4869-a16b-37d6030b2079","Type":"ContainerDied","Data":"0c3c968c83795ac1a7b40ac32ed075d700bd636eb4e9a6d04a4f18d663018271"} Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.147843 4770 scope.go:117] "RemoveContainer" containerID="8202959a2d6a9e3319d4f040fed6d702380bf5fbfc420149e5eda797c7472b95" Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.157238 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"93642b4e9136e576ce1ad898295dea2f212f632a6d746886f2e5f1af8035d45d"} Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.181143 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.488591741 podStartE2EDuration="5.181110406s" podCreationTimestamp="2025-10-04 03:24:28 +0000 UTC" firstStartedPulling="2025-10-04 03:24:28.96784347 +0000 UTC m=+1280.259853182" lastFinishedPulling="2025-10-04 03:24:32.660362135 +0000 UTC m=+1283.952371847" observedRunningTime="2025-10-04 03:24:33.168393203 +0000 UTC m=+1284.460402935" watchObservedRunningTime="2025-10-04 03:24:33.181110406 +0000 UTC m=+1284.473120118" Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.184365 4770 scope.go:117] "RemoveContainer" containerID="5a21c9e7067b28271f09914a49ea21aa7fa3dfb532c1ba2e54ec769996716b77" Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.222558 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-47pc5"] Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.235602 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d6d46f6cf-47pc5"] Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.306928 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4rvtr"] Oct 04 03:24:33 crc kubenswrapper[4770]: W1004 03:24:33.311347 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47a8258e_5ecc_44eb_8c4d_20ecd8078cb9.slice/crio-78bfa3be60af3528a366da235e8a8d57a0ade63270f357154b5515af7b9d3e96 WatchSource:0}: Error finding container 78bfa3be60af3528a366da235e8a8d57a0ade63270f357154b5515af7b9d3e96: Status 404 returned error can't find the container with id 78bfa3be60af3528a366da235e8a8d57a0ade63270f357154b5515af7b9d3e96 Oct 04 03:24:33 crc kubenswrapper[4770]: I1004 03:24:33.689318 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11acd15b-b791-4869-a16b-37d6030b2079" path="/var/lib/kubelet/pods/11acd15b-b791-4869-a16b-37d6030b2079/volumes" Oct 04 03:24:34 crc kubenswrapper[4770]: I1004 03:24:34.172739 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4rvtr" event={"ID":"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9","Type":"ContainerStarted","Data":"fdfbfac6fc788e30fc0d06f0b561b738e5d141700c8da1feb54964edb74a746c"} Oct 04 03:24:34 crc kubenswrapper[4770]: I1004 03:24:34.173052 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4rvtr" event={"ID":"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9","Type":"ContainerStarted","Data":"78bfa3be60af3528a366da235e8a8d57a0ade63270f357154b5515af7b9d3e96"} Oct 04 03:24:34 crc kubenswrapper[4770]: I1004 03:24:34.220544 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-4rvtr" podStartSLOduration=2.220510173 podStartE2EDuration="2.220510173s" podCreationTimestamp="2025-10-04 03:24:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:34.191676799 +0000 UTC m=+1285.483686511" watchObservedRunningTime="2025-10-04 03:24:34.220510173 +0000 UTC m=+1285.512520095" Oct 04 03:24:39 crc kubenswrapper[4770]: I1004 03:24:39.249481 4770 generic.go:334] "Generic (PLEG): container finished" podID="47a8258e-5ecc-44eb-8c4d-20ecd8078cb9" containerID="fdfbfac6fc788e30fc0d06f0b561b738e5d141700c8da1feb54964edb74a746c" exitCode=0 Oct 04 03:24:39 crc kubenswrapper[4770]: I1004 03:24:39.249605 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4rvtr" event={"ID":"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9","Type":"ContainerDied","Data":"fdfbfac6fc788e30fc0d06f0b561b738e5d141700c8da1feb54964edb74a746c"} Oct 04 03:24:39 crc kubenswrapper[4770]: I1004 03:24:39.430205 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 03:24:39 crc kubenswrapper[4770]: I1004 03:24:39.430323 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.446169 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.446213 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.779483 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.931296 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-scripts\") pod \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.931747 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-config-data\") pod \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.931800 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-combined-ca-bundle\") pod \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.931826 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvg4f\" (UniqueName: \"kubernetes.io/projected/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-kube-api-access-gvg4f\") pod \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\" (UID: \"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9\") " Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.945169 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-kube-api-access-gvg4f" (OuterVolumeSpecName: "kube-api-access-gvg4f") pod "47a8258e-5ecc-44eb-8c4d-20ecd8078cb9" (UID: "47a8258e-5ecc-44eb-8c4d-20ecd8078cb9"). InnerVolumeSpecName "kube-api-access-gvg4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.953812 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-scripts" (OuterVolumeSpecName: "scripts") pod "47a8258e-5ecc-44eb-8c4d-20ecd8078cb9" (UID: "47a8258e-5ecc-44eb-8c4d-20ecd8078cb9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.961206 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47a8258e-5ecc-44eb-8c4d-20ecd8078cb9" (UID: "47a8258e-5ecc-44eb-8c4d-20ecd8078cb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:40 crc kubenswrapper[4770]: I1004 03:24:40.974891 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-config-data" (OuterVolumeSpecName: "config-data") pod "47a8258e-5ecc-44eb-8c4d-20ecd8078cb9" (UID: "47a8258e-5ecc-44eb-8c4d-20ecd8078cb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.036065 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.036119 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.036137 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.036155 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvg4f\" (UniqueName: \"kubernetes.io/projected/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9-kube-api-access-gvg4f\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.271724 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4rvtr" event={"ID":"47a8258e-5ecc-44eb-8c4d-20ecd8078cb9","Type":"ContainerDied","Data":"78bfa3be60af3528a366da235e8a8d57a0ade63270f357154b5515af7b9d3e96"} Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.271797 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78bfa3be60af3528a366da235e8a8d57a0ade63270f357154b5515af7b9d3e96" Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.271897 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4rvtr" Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.493467 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.493735 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5a42a6c3-db14-49a3-bc09-79315b1c0169" containerName="nova-scheduler-scheduler" containerID="cri-o://8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1" gracePeriod=30 Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.510882 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.511571 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-log" containerID="cri-o://4df76b0965987281d66940d30bfcbf3da14ec8a918a94e04279e03639ad4bdaf" gracePeriod=30 Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.511734 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-api" containerID="cri-o://9fc4a0761d611105cc75c38fdc3d7d9340739f44d9017cc1b89853a831f1ee49" gracePeriod=30 Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.532073 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.532810 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-log" containerID="cri-o://9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee" gracePeriod=30 Oct 04 03:24:41 crc kubenswrapper[4770]: I1004 03:24:41.533556 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-metadata" containerID="cri-o://b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9" gracePeriod=30 Oct 04 03:24:42 crc kubenswrapper[4770]: I1004 03:24:42.285711 4770 generic.go:334] "Generic (PLEG): container finished" podID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerID="4df76b0965987281d66940d30bfcbf3da14ec8a918a94e04279e03639ad4bdaf" exitCode=143 Oct 04 03:24:42 crc kubenswrapper[4770]: I1004 03:24:42.286216 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd93c46b-fdd0-44ac-a503-673d32a72e27","Type":"ContainerDied","Data":"4df76b0965987281d66940d30bfcbf3da14ec8a918a94e04279e03639ad4bdaf"} Oct 04 03:24:42 crc kubenswrapper[4770]: I1004 03:24:42.289280 4770 generic.go:334] "Generic (PLEG): container finished" podID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerID="9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee" exitCode=143 Oct 04 03:24:42 crc kubenswrapper[4770]: I1004 03:24:42.289337 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0271b4e-6f09-49e2-9458-d495e7bf936a","Type":"ContainerDied","Data":"9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee"} Oct 04 03:24:43 crc kubenswrapper[4770]: E1004 03:24:43.959139 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1 is running failed: container process not found" containerID="8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 03:24:43 crc kubenswrapper[4770]: E1004 03:24:43.961143 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1 is running failed: container process not found" containerID="8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 03:24:43 crc kubenswrapper[4770]: E1004 03:24:43.961624 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1 is running failed: container process not found" containerID="8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 03:24:43 crc kubenswrapper[4770]: E1004 03:24:43.961686 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="5a42a6c3-db14-49a3-bc09-79315b1c0169" containerName="nova-scheduler-scheduler" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.240128 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.331621 4770 generic.go:334] "Generic (PLEG): container finished" podID="5a42a6c3-db14-49a3-bc09-79315b1c0169" containerID="8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1" exitCode=0 Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.331695 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a42a6c3-db14-49a3-bc09-79315b1c0169","Type":"ContainerDied","Data":"8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1"} Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.331744 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5a42a6c3-db14-49a3-bc09-79315b1c0169","Type":"ContainerDied","Data":"4670ac3ccfba07bfb82864f98d488b29544472d4e1c09be841dd2d46f6bfd803"} Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.331776 4770 scope.go:117] "RemoveContainer" containerID="8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.332029 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.366800 4770 scope.go:117] "RemoveContainer" containerID="8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1" Oct 04 03:24:44 crc kubenswrapper[4770]: E1004 03:24:44.367467 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1\": container with ID starting with 8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1 not found: ID does not exist" containerID="8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.367522 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1"} err="failed to get container status \"8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1\": rpc error: code = NotFound desc = could not find container \"8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1\": container with ID starting with 8108f804dbf71bce2be1c39220c66d5d19d0f39498334f0f202146a0d8dec5c1 not found: ID does not exist" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.412585 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-combined-ca-bundle\") pod \"5a42a6c3-db14-49a3-bc09-79315b1c0169\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.412946 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-config-data\") pod \"5a42a6c3-db14-49a3-bc09-79315b1c0169\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.413082 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr5qr\" (UniqueName: \"kubernetes.io/projected/5a42a6c3-db14-49a3-bc09-79315b1c0169-kube-api-access-vr5qr\") pod \"5a42a6c3-db14-49a3-bc09-79315b1c0169\" (UID: \"5a42a6c3-db14-49a3-bc09-79315b1c0169\") " Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.423942 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a42a6c3-db14-49a3-bc09-79315b1c0169-kube-api-access-vr5qr" (OuterVolumeSpecName: "kube-api-access-vr5qr") pod "5a42a6c3-db14-49a3-bc09-79315b1c0169" (UID: "5a42a6c3-db14-49a3-bc09-79315b1c0169"). InnerVolumeSpecName "kube-api-access-vr5qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.470260 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-config-data" (OuterVolumeSpecName: "config-data") pod "5a42a6c3-db14-49a3-bc09-79315b1c0169" (UID: "5a42a6c3-db14-49a3-bc09-79315b1c0169"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.470456 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a42a6c3-db14-49a3-bc09-79315b1c0169" (UID: "5a42a6c3-db14-49a3-bc09-79315b1c0169"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.515771 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.515825 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr5qr\" (UniqueName: \"kubernetes.io/projected/5a42a6c3-db14-49a3-bc09-79315b1c0169-kube-api-access-vr5qr\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.515839 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a42a6c3-db14-49a3-bc09-79315b1c0169-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.665871 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.675157 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.700513 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:24:44 crc kubenswrapper[4770]: E1004 03:24:44.701295 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11acd15b-b791-4869-a16b-37d6030b2079" containerName="dnsmasq-dns" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.701460 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="11acd15b-b791-4869-a16b-37d6030b2079" containerName="dnsmasq-dns" Oct 04 03:24:44 crc kubenswrapper[4770]: E1004 03:24:44.701568 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47a8258e-5ecc-44eb-8c4d-20ecd8078cb9" containerName="nova-manage" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.701648 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="47a8258e-5ecc-44eb-8c4d-20ecd8078cb9" containerName="nova-manage" Oct 04 03:24:44 crc kubenswrapper[4770]: E1004 03:24:44.701737 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11acd15b-b791-4869-a16b-37d6030b2079" containerName="init" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.701815 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="11acd15b-b791-4869-a16b-37d6030b2079" containerName="init" Oct 04 03:24:44 crc kubenswrapper[4770]: E1004 03:24:44.701877 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a42a6c3-db14-49a3-bc09-79315b1c0169" containerName="nova-scheduler-scheduler" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.701930 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a42a6c3-db14-49a3-bc09-79315b1c0169" containerName="nova-scheduler-scheduler" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.702186 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="11acd15b-b791-4869-a16b-37d6030b2079" containerName="dnsmasq-dns" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.702264 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a42a6c3-db14-49a3-bc09-79315b1c0169" containerName="nova-scheduler-scheduler" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.702333 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="47a8258e-5ecc-44eb-8c4d-20ecd8078cb9" containerName="nova-manage" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.703077 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.704756 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:36666->10.217.0.190:8775: read: connection reset by peer" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.704795 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:36676->10.217.0.190:8775: read: connection reset by peer" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.707393 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.713022 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.821197 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.821319 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-config-data\") pod \"nova-scheduler-0\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.821579 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngntj\" (UniqueName: \"kubernetes.io/projected/41a2c8f7-f9d1-44bf-8381-ceea9504a909-kube-api-access-ngntj\") pod \"nova-scheduler-0\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.923679 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-config-data\") pod \"nova-scheduler-0\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.923770 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngntj\" (UniqueName: \"kubernetes.io/projected/41a2c8f7-f9d1-44bf-8381-ceea9504a909-kube-api-access-ngntj\") pod \"nova-scheduler-0\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.923892 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.930430 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-config-data\") pod \"nova-scheduler-0\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.932930 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " pod="openstack/nova-scheduler-0" Oct 04 03:24:44 crc kubenswrapper[4770]: I1004 03:24:44.944957 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngntj\" (UniqueName: \"kubernetes.io/projected/41a2c8f7-f9d1-44bf-8381-ceea9504a909-kube-api-access-ngntj\") pod \"nova-scheduler-0\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " pod="openstack/nova-scheduler-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.093643 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.132619 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.229803 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-config-data\") pod \"b0271b4e-6f09-49e2-9458-d495e7bf936a\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.229962 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-nova-metadata-tls-certs\") pod \"b0271b4e-6f09-49e2-9458-d495e7bf936a\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.230208 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77l7w\" (UniqueName: \"kubernetes.io/projected/b0271b4e-6f09-49e2-9458-d495e7bf936a-kube-api-access-77l7w\") pod \"b0271b4e-6f09-49e2-9458-d495e7bf936a\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.230275 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-combined-ca-bundle\") pod \"b0271b4e-6f09-49e2-9458-d495e7bf936a\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.230302 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0271b4e-6f09-49e2-9458-d495e7bf936a-logs\") pod \"b0271b4e-6f09-49e2-9458-d495e7bf936a\" (UID: \"b0271b4e-6f09-49e2-9458-d495e7bf936a\") " Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.231413 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0271b4e-6f09-49e2-9458-d495e7bf936a-logs" (OuterVolumeSpecName: "logs") pod "b0271b4e-6f09-49e2-9458-d495e7bf936a" (UID: "b0271b4e-6f09-49e2-9458-d495e7bf936a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.242692 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0271b4e-6f09-49e2-9458-d495e7bf936a-kube-api-access-77l7w" (OuterVolumeSpecName: "kube-api-access-77l7w") pod "b0271b4e-6f09-49e2-9458-d495e7bf936a" (UID: "b0271b4e-6f09-49e2-9458-d495e7bf936a"). InnerVolumeSpecName "kube-api-access-77l7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.269598 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-config-data" (OuterVolumeSpecName: "config-data") pod "b0271b4e-6f09-49e2-9458-d495e7bf936a" (UID: "b0271b4e-6f09-49e2-9458-d495e7bf936a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.269627 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0271b4e-6f09-49e2-9458-d495e7bf936a" (UID: "b0271b4e-6f09-49e2-9458-d495e7bf936a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.296361 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b0271b4e-6f09-49e2-9458-d495e7bf936a" (UID: "b0271b4e-6f09-49e2-9458-d495e7bf936a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.333282 4770 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.333319 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77l7w\" (UniqueName: \"kubernetes.io/projected/b0271b4e-6f09-49e2-9458-d495e7bf936a-kube-api-access-77l7w\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.333332 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.333343 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0271b4e-6f09-49e2-9458-d495e7bf936a-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.333353 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0271b4e-6f09-49e2-9458-d495e7bf936a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.343844 4770 generic.go:334] "Generic (PLEG): container finished" podID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerID="b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9" exitCode=0 Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.343903 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0271b4e-6f09-49e2-9458-d495e7bf936a","Type":"ContainerDied","Data":"b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9"} Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.343942 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0271b4e-6f09-49e2-9458-d495e7bf936a","Type":"ContainerDied","Data":"6ca0e1685f3d9eb5381783c9fa1710ec56c6c9c973345194ac0da66eed338d54"} Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.343965 4770 scope.go:117] "RemoveContainer" containerID="b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.344300 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.398024 4770 scope.go:117] "RemoveContainer" containerID="9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.398113 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.406977 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.434991 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:24:45 crc kubenswrapper[4770]: E1004 03:24:45.435539 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-log" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.435568 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-log" Oct 04 03:24:45 crc kubenswrapper[4770]: E1004 03:24:45.435704 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-metadata" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.435717 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-metadata" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.435964 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-log" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.435998 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" containerName="nova-metadata-metadata" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.437764 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.440742 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.440910 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.444324 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.461859 4770 scope.go:117] "RemoveContainer" containerID="b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9" Oct 04 03:24:45 crc kubenswrapper[4770]: E1004 03:24:45.462903 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9\": container with ID starting with b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9 not found: ID does not exist" containerID="b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.462943 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9"} err="failed to get container status \"b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9\": rpc error: code = NotFound desc = could not find container \"b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9\": container with ID starting with b40f12800d12e733fcbec4b45fdde593a8ee8f02d71a5ae7c54c270a2a3990a9 not found: ID does not exist" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.462973 4770 scope.go:117] "RemoveContainer" containerID="9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee" Oct 04 03:24:45 crc kubenswrapper[4770]: E1004 03:24:45.465415 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee\": container with ID starting with 9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee not found: ID does not exist" containerID="9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.465471 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee"} err="failed to get container status \"9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee\": rpc error: code = NotFound desc = could not find container \"9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee\": container with ID starting with 9defbf30c6ebb78662c37a20270e56be9895015682069d9d925a6f9a31a534ee not found: ID does not exist" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.536692 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.536803 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nppcd\" (UniqueName: \"kubernetes.io/projected/79ab7038-39be-4c09-82ca-00f83aa69e37-kube-api-access-nppcd\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.536848 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-config-data\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.536874 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.536925 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ab7038-39be-4c09-82ca-00f83aa69e37-logs\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.598873 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:24:45 crc kubenswrapper[4770]: W1004 03:24:45.599457 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41a2c8f7_f9d1_44bf_8381_ceea9504a909.slice/crio-769613dedd149600ee44a38e77feeb871c18558cd7951f9032fbd9ff8812e13b WatchSource:0}: Error finding container 769613dedd149600ee44a38e77feeb871c18558cd7951f9032fbd9ff8812e13b: Status 404 returned error can't find the container with id 769613dedd149600ee44a38e77feeb871c18558cd7951f9032fbd9ff8812e13b Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.638357 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-config-data\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.638419 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.638476 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ab7038-39be-4c09-82ca-00f83aa69e37-logs\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.638546 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.638618 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nppcd\" (UniqueName: \"kubernetes.io/projected/79ab7038-39be-4c09-82ca-00f83aa69e37-kube-api-access-nppcd\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.639244 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ab7038-39be-4c09-82ca-00f83aa69e37-logs\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.644717 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.644766 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.645206 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-config-data\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.657042 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nppcd\" (UniqueName: \"kubernetes.io/projected/79ab7038-39be-4c09-82ca-00f83aa69e37-kube-api-access-nppcd\") pod \"nova-metadata-0\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " pod="openstack/nova-metadata-0" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.692318 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a42a6c3-db14-49a3-bc09-79315b1c0169" path="/var/lib/kubelet/pods/5a42a6c3-db14-49a3-bc09-79315b1c0169/volumes" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.693332 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0271b4e-6f09-49e2-9458-d495e7bf936a" path="/var/lib/kubelet/pods/b0271b4e-6f09-49e2-9458-d495e7bf936a/volumes" Oct 04 03:24:45 crc kubenswrapper[4770]: I1004 03:24:45.763721 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:24:46 crc kubenswrapper[4770]: I1004 03:24:46.357146 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd93c46b-fdd0-44ac-a503-673d32a72e27","Type":"ContainerDied","Data":"9fc4a0761d611105cc75c38fdc3d7d9340739f44d9017cc1b89853a831f1ee49"} Oct 04 03:24:46 crc kubenswrapper[4770]: I1004 03:24:46.357184 4770 generic.go:334] "Generic (PLEG): container finished" podID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerID="9fc4a0761d611105cc75c38fdc3d7d9340739f44d9017cc1b89853a831f1ee49" exitCode=0 Oct 04 03:24:46 crc kubenswrapper[4770]: I1004 03:24:46.361812 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"41a2c8f7-f9d1-44bf-8381-ceea9504a909","Type":"ContainerStarted","Data":"c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa"} Oct 04 03:24:46 crc kubenswrapper[4770]: I1004 03:24:46.361879 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"41a2c8f7-f9d1-44bf-8381-ceea9504a909","Type":"ContainerStarted","Data":"769613dedd149600ee44a38e77feeb871c18558cd7951f9032fbd9ff8812e13b"} Oct 04 03:24:46 crc kubenswrapper[4770]: I1004 03:24:46.401995 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.401966151 podStartE2EDuration="2.401966151s" podCreationTimestamp="2025-10-04 03:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:46.385690816 +0000 UTC m=+1297.677700528" watchObservedRunningTime="2025-10-04 03:24:46.401966151 +0000 UTC m=+1297.693975893" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.065813 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.067501 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.188756 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd93c46b-fdd0-44ac-a503-673d32a72e27-logs\") pod \"dd93c46b-fdd0-44ac-a503-673d32a72e27\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.188804 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-config-data\") pod \"dd93c46b-fdd0-44ac-a503-673d32a72e27\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.188925 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-public-tls-certs\") pod \"dd93c46b-fdd0-44ac-a503-673d32a72e27\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.188963 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-internal-tls-certs\") pod \"dd93c46b-fdd0-44ac-a503-673d32a72e27\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.188993 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-combined-ca-bundle\") pod \"dd93c46b-fdd0-44ac-a503-673d32a72e27\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.189062 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbfnr\" (UniqueName: \"kubernetes.io/projected/dd93c46b-fdd0-44ac-a503-673d32a72e27-kube-api-access-gbfnr\") pod \"dd93c46b-fdd0-44ac-a503-673d32a72e27\" (UID: \"dd93c46b-fdd0-44ac-a503-673d32a72e27\") " Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.192270 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd93c46b-fdd0-44ac-a503-673d32a72e27-logs" (OuterVolumeSpecName: "logs") pod "dd93c46b-fdd0-44ac-a503-673d32a72e27" (UID: "dd93c46b-fdd0-44ac-a503-673d32a72e27"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.199479 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd93c46b-fdd0-44ac-a503-673d32a72e27-kube-api-access-gbfnr" (OuterVolumeSpecName: "kube-api-access-gbfnr") pod "dd93c46b-fdd0-44ac-a503-673d32a72e27" (UID: "dd93c46b-fdd0-44ac-a503-673d32a72e27"). InnerVolumeSpecName "kube-api-access-gbfnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.263876 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd93c46b-fdd0-44ac-a503-673d32a72e27" (UID: "dd93c46b-fdd0-44ac-a503-673d32a72e27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.269117 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-config-data" (OuterVolumeSpecName: "config-data") pod "dd93c46b-fdd0-44ac-a503-673d32a72e27" (UID: "dd93c46b-fdd0-44ac-a503-673d32a72e27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.291504 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.291537 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.291550 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbfnr\" (UniqueName: \"kubernetes.io/projected/dd93c46b-fdd0-44ac-a503-673d32a72e27-kube-api-access-gbfnr\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.291558 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd93c46b-fdd0-44ac-a503-673d32a72e27-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.292474 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "dd93c46b-fdd0-44ac-a503-673d32a72e27" (UID: "dd93c46b-fdd0-44ac-a503-673d32a72e27"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.310305 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dd93c46b-fdd0-44ac-a503-673d32a72e27" (UID: "dd93c46b-fdd0-44ac-a503-673d32a72e27"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.370685 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79ab7038-39be-4c09-82ca-00f83aa69e37","Type":"ContainerStarted","Data":"1e4a44749d34e20093fa0ef2448d2cdc6233bb16e1aa47b08bdfb010277a16ae"} Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.370745 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79ab7038-39be-4c09-82ca-00f83aa69e37","Type":"ContainerStarted","Data":"bdec7769a483ec499ca82ac6def167d67b14c965f9ed69c0964ade9fd5a29f1b"} Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.372698 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dd93c46b-fdd0-44ac-a503-673d32a72e27","Type":"ContainerDied","Data":"f824d739053c047ad163b5150e861626f83dfb1106f93593c99452b944bed303"} Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.372740 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.372749 4770 scope.go:117] "RemoveContainer" containerID="9fc4a0761d611105cc75c38fdc3d7d9340739f44d9017cc1b89853a831f1ee49" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.400078 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.400116 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd93c46b-fdd0-44ac-a503-673d32a72e27-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.406103 4770 scope.go:117] "RemoveContainer" containerID="4df76b0965987281d66940d30bfcbf3da14ec8a918a94e04279e03639ad4bdaf" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.412087 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.421868 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.439297 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:47 crc kubenswrapper[4770]: E1004 03:24:47.439690 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-api" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.439709 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-api" Oct 04 03:24:47 crc kubenswrapper[4770]: E1004 03:24:47.439751 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-log" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.439759 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-log" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.439942 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-api" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.439967 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" containerName="nova-api-log" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.441842 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.446606 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.447054 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.447418 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.475103 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.604347 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.604534 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98pqk\" (UniqueName: \"kubernetes.io/projected/4b795408-0144-4d73-8a56-043ce3e38ee6-kube-api-access-98pqk\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.604726 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-config-data\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.604782 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b795408-0144-4d73-8a56-043ce3e38ee6-logs\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.604839 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.605334 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-public-tls-certs\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.685865 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd93c46b-fdd0-44ac-a503-673d32a72e27" path="/var/lib/kubelet/pods/dd93c46b-fdd0-44ac-a503-673d32a72e27/volumes" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.707520 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-config-data\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.707582 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b795408-0144-4d73-8a56-043ce3e38ee6-logs\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.707607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.707726 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-public-tls-certs\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.707779 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.707809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98pqk\" (UniqueName: \"kubernetes.io/projected/4b795408-0144-4d73-8a56-043ce3e38ee6-kube-api-access-98pqk\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.708187 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b795408-0144-4d73-8a56-043ce3e38ee6-logs\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.714834 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-config-data\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.714937 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.716483 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.716615 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-public-tls-certs\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.724987 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98pqk\" (UniqueName: \"kubernetes.io/projected/4b795408-0144-4d73-8a56-043ce3e38ee6-kube-api-access-98pqk\") pod \"nova-api-0\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " pod="openstack/nova-api-0" Oct 04 03:24:47 crc kubenswrapper[4770]: I1004 03:24:47.784063 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:24:48 crc kubenswrapper[4770]: I1004 03:24:48.267311 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:24:48 crc kubenswrapper[4770]: W1004 03:24:48.285573 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b795408_0144_4d73_8a56_043ce3e38ee6.slice/crio-c686cdba7b99d359ef3efed5230a52bf695646bcd700ca78bdc38e78274cf6f7 WatchSource:0}: Error finding container c686cdba7b99d359ef3efed5230a52bf695646bcd700ca78bdc38e78274cf6f7: Status 404 returned error can't find the container with id c686cdba7b99d359ef3efed5230a52bf695646bcd700ca78bdc38e78274cf6f7 Oct 04 03:24:48 crc kubenswrapper[4770]: I1004 03:24:48.383219 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b795408-0144-4d73-8a56-043ce3e38ee6","Type":"ContainerStarted","Data":"c686cdba7b99d359ef3efed5230a52bf695646bcd700ca78bdc38e78274cf6f7"} Oct 04 03:24:48 crc kubenswrapper[4770]: I1004 03:24:48.386335 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79ab7038-39be-4c09-82ca-00f83aa69e37","Type":"ContainerStarted","Data":"fa7c4e0199de7a74a6a51a933f598585e135eac2ee7b99b1d3d49fa44c55ad5a"} Oct 04 03:24:48 crc kubenswrapper[4770]: I1004 03:24:48.408644 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.408623729 podStartE2EDuration="3.408623729s" podCreationTimestamp="2025-10-04 03:24:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:48.408328822 +0000 UTC m=+1299.700338544" watchObservedRunningTime="2025-10-04 03:24:48.408623729 +0000 UTC m=+1299.700633441" Oct 04 03:24:49 crc kubenswrapper[4770]: I1004 03:24:49.404282 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b795408-0144-4d73-8a56-043ce3e38ee6","Type":"ContainerStarted","Data":"6bba8b83577d71e684e12c6e57b942c931915975b422c9aa9892702880a260fc"} Oct 04 03:24:49 crc kubenswrapper[4770]: I1004 03:24:49.404548 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b795408-0144-4d73-8a56-043ce3e38ee6","Type":"ContainerStarted","Data":"30a450deeec6b43046f1e4c12cadb70ffef33664a48ce36b69bf552dadbd53c4"} Oct 04 03:24:49 crc kubenswrapper[4770]: I1004 03:24:49.434454 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.434429221 podStartE2EDuration="2.434429221s" podCreationTimestamp="2025-10-04 03:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:24:49.432556293 +0000 UTC m=+1300.724566035" watchObservedRunningTime="2025-10-04 03:24:49.434429221 +0000 UTC m=+1300.726438963" Oct 04 03:24:50 crc kubenswrapper[4770]: I1004 03:24:50.095647 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 03:24:50 crc kubenswrapper[4770]: I1004 03:24:50.764199 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 03:24:50 crc kubenswrapper[4770]: I1004 03:24:50.764264 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 03:24:55 crc kubenswrapper[4770]: I1004 03:24:55.095311 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 03:24:55 crc kubenswrapper[4770]: I1004 03:24:55.142516 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 03:24:55 crc kubenswrapper[4770]: I1004 03:24:55.505413 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 03:24:55 crc kubenswrapper[4770]: I1004 03:24:55.764835 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 03:24:55 crc kubenswrapper[4770]: I1004 03:24:55.764903 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 03:24:56 crc kubenswrapper[4770]: I1004 03:24:56.847373 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:56 crc kubenswrapper[4770]: I1004 03:24:56.847422 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:57 crc kubenswrapper[4770]: I1004 03:24:57.784808 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 03:24:57 crc kubenswrapper[4770]: I1004 03:24:57.784889 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 03:24:58 crc kubenswrapper[4770]: I1004 03:24:58.456737 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 03:24:58 crc kubenswrapper[4770]: I1004 03:24:58.803270 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 03:24:58 crc kubenswrapper[4770]: I1004 03:24:58.803353 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 03:25:05 crc kubenswrapper[4770]: I1004 03:25:05.772487 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 03:25:05 crc kubenswrapper[4770]: I1004 03:25:05.773723 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 03:25:05 crc kubenswrapper[4770]: I1004 03:25:05.783052 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 03:25:06 crc kubenswrapper[4770]: I1004 03:25:06.611108 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 03:25:07 crc kubenswrapper[4770]: I1004 03:25:07.803801 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 03:25:07 crc kubenswrapper[4770]: I1004 03:25:07.804434 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 03:25:07 crc kubenswrapper[4770]: I1004 03:25:07.806631 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 03:25:07 crc kubenswrapper[4770]: I1004 03:25:07.812126 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 03:25:08 crc kubenswrapper[4770]: I1004 03:25:08.631855 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 03:25:08 crc kubenswrapper[4770]: I1004 03:25:08.640873 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.632684 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-576495ccf7-bslzp"] Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.635050 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.673092 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-867bd5444b-xx8d4"] Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.674825 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.696986 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-576495ccf7-bslzp"] Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.733679 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-867bd5444b-xx8d4"] Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.802645 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8mms\" (UniqueName: \"kubernetes.io/projected/9b1ac142-741d-48a1-bf5c-982bbf7bace6-kube-api-access-d8mms\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.802703 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.802774 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w59hs\" (UniqueName: \"kubernetes.io/projected/8207573d-3535-4814-aff9-15c30c7600c7-kube-api-access-w59hs\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.802795 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8207573d-3535-4814-aff9-15c30c7600c7-logs\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.802831 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.802886 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-combined-ca-bundle\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.802901 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-combined-ca-bundle\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.802920 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data-custom\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.803046 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b1ac142-741d-48a1-bf5c-982bbf7bace6-logs\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.803070 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.860432 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.897703 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.898294 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="5d67ab7f-c0b5-476c-b5d2-99b90eed466a" containerName="openstackclient" containerID="cri-o://f4087f933afeeb3b260782f8334f21234028523075122f703a71f7c541bb19dc" gracePeriod=2 Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.905437 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8mms\" (UniqueName: \"kubernetes.io/projected/9b1ac142-741d-48a1-bf5c-982bbf7bace6-kube-api-access-d8mms\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.905681 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.905839 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w59hs\" (UniqueName: \"kubernetes.io/projected/8207573d-3535-4814-aff9-15c30c7600c7-kube-api-access-w59hs\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.905994 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8207573d-3535-4814-aff9-15c30c7600c7-logs\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.906141 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.906262 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-combined-ca-bundle\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.906427 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data-custom\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.906514 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-combined-ca-bundle\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.906769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b1ac142-741d-48a1-bf5c-982bbf7bace6-logs\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.906880 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.909109 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8207573d-3535-4814-aff9-15c30c7600c7-logs\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.913726 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b1ac142-741d-48a1-bf5c-982bbf7bace6-logs\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.943700 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.950566 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-combined-ca-bundle\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.953025 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.962954 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.963992 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="57f70627-2bdd-4780-8044-75c3de4aee05" containerName="openstack-network-exporter" containerID="cri-o://760e8cc98a032092f27104dafba09f778ecc47ca79026332349329b345264e45" gracePeriod=300 Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.972247 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w59hs\" (UniqueName: \"kubernetes.io/projected/8207573d-3535-4814-aff9-15c30c7600c7-kube-api-access-w59hs\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.972269 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-combined-ca-bundle\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.972465 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.972345 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data-custom\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.972994 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data\") pod \"barbican-worker-576495ccf7-bslzp\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.993549 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.994158 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="ovn-northd" containerID="cri-o://38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" gracePeriod=30 Oct 04 03:25:30 crc kubenswrapper[4770]: I1004 03:25:30.994387 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="openstack-network-exporter" containerID="cri-o://00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90" gracePeriod=30 Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.025577 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:25:31 crc kubenswrapper[4770]: E1004 03:25:31.027979 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 04 03:25:31 crc kubenswrapper[4770]: E1004 03:25:31.028143 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data podName:2def82a8-5927-4ebe-ac87-e8ad232797ee nodeName:}" failed. No retries permitted until 2025-10-04 03:25:31.528124622 +0000 UTC m=+1342.820134334 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data") pod "rabbitmq-server-0" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee") : configmap "rabbitmq-config-data" not found Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.069307 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinderec07-account-delete-j44sn"] Oct 04 03:25:31 crc kubenswrapper[4770]: E1004 03:25:31.069961 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d67ab7f-c0b5-476c-b5d2-99b90eed466a" containerName="openstackclient" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.069972 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d67ab7f-c0b5-476c-b5d2-99b90eed466a" containerName="openstackclient" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.070206 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d67ab7f-c0b5-476c-b5d2-99b90eed466a" containerName="openstackclient" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.071032 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderec07-account-delete-j44sn" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.089494 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8mms\" (UniqueName: \"kubernetes.io/projected/9b1ac142-741d-48a1-bf5c-982bbf7bace6-kube-api-access-d8mms\") pod \"barbican-keystone-listener-867bd5444b-xx8d4\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.125579 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgnqx\" (UniqueName: \"kubernetes.io/projected/2af9166c-e2c5-462d-be52-340be8aadfd5-kube-api-access-wgnqx\") pod \"cinderec07-account-delete-j44sn\" (UID: \"2af9166c-e2c5-462d-be52-340be8aadfd5\") " pod="openstack/cinderec07-account-delete-j44sn" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.125828 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderec07-account-delete-j44sn"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.162094 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.162978 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerName="openstack-network-exporter" containerID="cri-o://347a7235cf642c5bee0dc831171a48e6731de85a58d10232d67b2b94de43b821" gracePeriod=300 Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.184886 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-hkqts"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.214040 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-hkqts"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.227540 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgnqx\" (UniqueName: \"kubernetes.io/projected/2af9166c-e2c5-462d-be52-340be8aadfd5-kube-api-access-wgnqx\") pod \"cinderec07-account-delete-j44sn\" (UID: \"2af9166c-e2c5-462d-be52-340be8aadfd5\") " pod="openstack/cinderec07-account-delete-j44sn" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.244763 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="57f70627-2bdd-4780-8044-75c3de4aee05" containerName="ovsdbserver-nb" containerID="cri-o://92885b43eda03ead2301755a74384e53d3b27a5aa4fddd4305ff2576a9f59a7c" gracePeriod=300 Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.250523 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutronb0c9-account-delete-hc97s"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.268297 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb0c9-account-delete-hc97s" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.287263 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronb0c9-account-delete-hc97s"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.297218 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgnqx\" (UniqueName: \"kubernetes.io/projected/2af9166c-e2c5-462d-be52-340be8aadfd5-kube-api-access-wgnqx\") pod \"cinderec07-account-delete-j44sn\" (UID: \"2af9166c-e2c5-462d-be52-340be8aadfd5\") " pod="openstack/cinderec07-account-delete-j44sn" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.319571 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderec07-account-delete-j44sn" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.322052 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementded7-account-delete-qxzt9"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.324189 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementded7-account-delete-qxzt9" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.364957 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.400538 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerName="ovsdbserver-sb" containerID="cri-o://ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a" gracePeriod=300 Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.418544 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-hj2s4"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.433354 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzx4s\" (UniqueName: \"kubernetes.io/projected/d18c6351-6663-4a3c-84b6-2752f5e66e49-kube-api-access-kzx4s\") pod \"placementded7-account-delete-qxzt9\" (UID: \"d18c6351-6663-4a3c-84b6-2752f5e66e49\") " pod="openstack/placementded7-account-delete-qxzt9" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.433417 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nblt4\" (UniqueName: \"kubernetes.io/projected/c00a8476-9758-46f5-864f-9a7b80ca8bd1-kube-api-access-nblt4\") pod \"neutronb0c9-account-delete-hc97s\" (UID: \"c00a8476-9758-46f5-864f-9a7b80ca8bd1\") " pod="openstack/neutronb0c9-account-delete-hc97s" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.436203 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-hj2s4"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.450750 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementded7-account-delete-qxzt9"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.468873 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-6nz6q"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.539343 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzx4s\" (UniqueName: \"kubernetes.io/projected/d18c6351-6663-4a3c-84b6-2752f5e66e49-kube-api-access-kzx4s\") pod \"placementded7-account-delete-qxzt9\" (UID: \"d18c6351-6663-4a3c-84b6-2752f5e66e49\") " pod="openstack/placementded7-account-delete-qxzt9" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.539510 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nblt4\" (UniqueName: \"kubernetes.io/projected/c00a8476-9758-46f5-864f-9a7b80ca8bd1-kube-api-access-nblt4\") pod \"neutronb0c9-account-delete-hc97s\" (UID: \"c00a8476-9758-46f5-864f-9a7b80ca8bd1\") " pod="openstack/neutronb0c9-account-delete-hc97s" Oct 04 03:25:31 crc kubenswrapper[4770]: E1004 03:25:31.540268 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 04 03:25:31 crc kubenswrapper[4770]: E1004 03:25:31.540351 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data podName:2def82a8-5927-4ebe-ac87-e8ad232797ee nodeName:}" failed. No retries permitted until 2025-10-04 03:25:32.54032255 +0000 UTC m=+1343.832332262 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data") pod "rabbitmq-server-0" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee") : configmap "rabbitmq-config-data" not found Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.565682 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glanced8ca-account-delete-k2b6r"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.579569 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glanced8ca-account-delete-k2b6r" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.596269 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nblt4\" (UniqueName: \"kubernetes.io/projected/c00a8476-9758-46f5-864f-9a7b80ca8bd1-kube-api-access-nblt4\") pod \"neutronb0c9-account-delete-hc97s\" (UID: \"c00a8476-9758-46f5-864f-9a7b80ca8bd1\") " pod="openstack/neutronb0c9-account-delete-hc97s" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.642833 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzx4s\" (UniqueName: \"kubernetes.io/projected/d18c6351-6663-4a3c-84b6-2752f5e66e49-kube-api-access-kzx4s\") pod \"placementded7-account-delete-qxzt9\" (UID: \"d18c6351-6663-4a3c-84b6-2752f5e66e49\") " pod="openstack/placementded7-account-delete-qxzt9" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.643888 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb0c9-account-delete-hc97s" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.671529 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-c2v9x"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.739184 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a8836e2-0ea8-43cb-ba3f-535b106c4f3a" path="/var/lib/kubelet/pods/3a8836e2-0ea8-43cb-ba3f-535b106c4f3a/volumes" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.740247 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b736169c-f680-43f5-b088-110ec7e664d8" path="/var/lib/kubelet/pods/b736169c-f680-43f5-b088-110ec7e664d8/volumes" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.750206 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjgk4\" (UniqueName: \"kubernetes.io/projected/d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c-kube-api-access-qjgk4\") pod \"glanced8ca-account-delete-k2b6r\" (UID: \"d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c\") " pod="openstack/glanced8ca-account-delete-k2b6r" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.766605 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glanced8ca-account-delete-k2b6r"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.750507 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementded7-account-delete-qxzt9" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.814137 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-tn765"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.814389 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-tn765" podUID="b4444038-6604-4e51-b27d-20d3616bf640" containerName="openstack-network-exporter" containerID="cri-o://3089621a8e69e00ac3e8841ec02e869f55aaa2a7703a475f407db0d736035cd6" gracePeriod=30 Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.859560 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjgk4\" (UniqueName: \"kubernetes.io/projected/d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c-kube-api-access-qjgk4\") pod \"glanced8ca-account-delete-k2b6r\" (UID: \"d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c\") " pod="openstack/glanced8ca-account-delete-k2b6r" Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.869871 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.910910 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6886d"] Oct 04 03:25:31 crc kubenswrapper[4770]: I1004 03:25:31.946198 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6886d"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.012872 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjgk4\" (UniqueName: \"kubernetes.io/projected/d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c-kube-api-access-qjgk4\") pod \"glanced8ca-account-delete-k2b6r\" (UID: \"d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c\") " pod="openstack/glanced8ca-account-delete-k2b6r" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.013543 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_57f70627-2bdd-4780-8044-75c3de4aee05/ovsdbserver-nb/0.log" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.013658 4770 generic.go:334] "Generic (PLEG): container finished" podID="57f70627-2bdd-4780-8044-75c3de4aee05" containerID="760e8cc98a032092f27104dafba09f778ecc47ca79026332349329b345264e45" exitCode=2 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.013724 4770 generic.go:334] "Generic (PLEG): container finished" podID="57f70627-2bdd-4780-8044-75c3de4aee05" containerID="92885b43eda03ead2301755a74384e53d3b27a5aa4fddd4305ff2576a9f59a7c" exitCode=143 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.013851 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"57f70627-2bdd-4780-8044-75c3de4aee05","Type":"ContainerDied","Data":"760e8cc98a032092f27104dafba09f778ecc47ca79026332349329b345264e45"} Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.013937 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"57f70627-2bdd-4780-8044-75c3de4aee05","Type":"ContainerDied","Data":"92885b43eda03ead2301755a74384e53d3b27a5aa4fddd4305ff2576a9f59a7c"} Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.028027 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi3321-account-delete-gmw9l"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.029486 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3321-account-delete-gmw9l" Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.066992 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.067089 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data podName:642fe06d-e5b8-461e-a1f7-dbcefc071945 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:32.567070748 +0000 UTC m=+1343.859080460 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data") pod "rabbitmq-cell1-server-0" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945") : configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.068618 4770 generic.go:334] "Generic (PLEG): container finished" podID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerID="00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90" exitCode=2 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.068795 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a36939a0-d10b-488e-8bd4-165b1f9ae4e9","Type":"ContainerDied","Data":"00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90"} Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.073062 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi3321-account-delete-gmw9l"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.106821 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d4508feb-4ff8-47ec-86d7-7a9a0ad61a24/ovsdbserver-sb/0.log" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.106869 4770 generic.go:334] "Generic (PLEG): container finished" podID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerID="347a7235cf642c5bee0dc831171a48e6731de85a58d10232d67b2b94de43b821" exitCode=2 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.106888 4770 generic.go:334] "Generic (PLEG): container finished" podID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerID="ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a" exitCode=143 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.106908 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24","Type":"ContainerDied","Data":"347a7235cf642c5bee0dc831171a48e6731de85a58d10232d67b2b94de43b821"} Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.106938 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24","Type":"ContainerDied","Data":"ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a"} Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.175068 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell109d5-account-delete-6g4cl"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.176467 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell109d5-account-delete-6g4cl" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.192394 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbttb\" (UniqueName: \"kubernetes.io/projected/6d5328f9-e5e9-4ad5-96ff-02442117ff0e-kube-api-access-jbttb\") pod \"novaapi3321-account-delete-gmw9l\" (UID: \"6d5328f9-e5e9-4ad5-96ff-02442117ff0e\") " pod="openstack/novaapi3321-account-delete-gmw9l" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.276510 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glanced8ca-account-delete-k2b6r" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.287090 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell109d5-account-delete-6g4cl"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.310380 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbttb\" (UniqueName: \"kubernetes.io/projected/6d5328f9-e5e9-4ad5-96ff-02442117ff0e-kube-api-access-jbttb\") pod \"novaapi3321-account-delete-gmw9l\" (UID: \"6d5328f9-e5e9-4ad5-96ff-02442117ff0e\") " pod="openstack/novaapi3321-account-delete-gmw9l" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.310422 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhxhj\" (UniqueName: \"kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj\") pod \"novacell109d5-account-delete-6g4cl\" (UID: \"0f56d0a9-6e93-47b0-8ba2-4b04020f9296\") " pod="openstack/novacell109d5-account-delete-6g4cl" Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.352893 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.356161 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-ql7kt"] Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.368164 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.399611 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-ql7kt"] Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.402195 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.402607 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="ovn-northd" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.403247 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbttb\" (UniqueName: \"kubernetes.io/projected/6d5328f9-e5e9-4ad5-96ff-02442117ff0e-kube-api-access-jbttb\") pod \"novaapi3321-account-delete-gmw9l\" (UID: \"6d5328f9-e5e9-4ad5-96ff-02442117ff0e\") " pod="openstack/novaapi3321-account-delete-gmw9l" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.415515 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhxhj\" (UniqueName: \"kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj\") pod \"novacell109d5-account-delete-6g4cl\" (UID: \"0f56d0a9-6e93-47b0-8ba2-4b04020f9296\") " pod="openstack/novacell109d5-account-delete-6g4cl" Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.431976 4770 projected.go:194] Error preparing data for projected volume kube-api-access-qhxhj for pod openstack/novacell109d5-account-delete-6g4cl: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.432091 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj podName:0f56d0a9-6e93-47b0-8ba2-4b04020f9296 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:32.932069975 +0000 UTC m=+1344.224079688 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-qhxhj" (UniqueName: "kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj") pod "novacell109d5-account-delete-6g4cl" (UID: "0f56d0a9-6e93-47b0-8ba2-4b04020f9296") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.439857 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-6qh86"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.482946 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3321-account-delete-gmw9l" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.493254 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.494625 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerName="cinder-scheduler" containerID="cri-o://24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.495082 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerName="probe" containerID="cri-o://1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.529529 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-6qh86"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.558890 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.565163 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" containerName="cinder-api-log" containerID="cri-o://68b1223ff961d5213ea6bc5690e64d21567c84142a78aaf580e2d2077be0f0d2" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.565444 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" containerName="cinder-api" containerID="cri-o://2ba074cd36ae8e2080bb5c06e842e5c1124a1cecb16c5a3f5f1ebd17d268f918" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.625559 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.625626 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data podName:642fe06d-e5b8-461e-a1f7-dbcefc071945 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:33.625607758 +0000 UTC m=+1344.917617470 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data") pod "rabbitmq-cell1-server-0" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945") : configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.628516 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.628601 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data podName:2def82a8-5927-4ebe-ac87-e8ad232797ee nodeName:}" failed. No retries permitted until 2025-10-04 03:25:34.628578095 +0000 UTC m=+1345.920587807 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data") pod "rabbitmq-server-0" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee") : configmap "rabbitmq-config-data" not found Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.641461 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-bl2rc"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.661571 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662027 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-server" containerID="cri-o://f824e75e4b4bc8b5e89c5df2fd7e73c49c2760133924bc91336e17d29f37ffa6" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662512 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-updater" containerID="cri-o://0bbc28ed4f22995dd014ebc4247f63d2b41fd2841242ce9db1f8c81ed20dd754" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662655 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="swift-recon-cron" containerID="cri-o://345b8c0092ebcc04825e922384f0c8ee2e508d97ba89f66363c61e92ce3d2670" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662696 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="rsync" containerID="cri-o://2aea2997649b938d2836a8f46cd25d7caab6d10e8f2aeeac5509198aa88b89ca" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662747 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-expirer" containerID="cri-o://6385e85337f2e8e8a6703c8fe588bbfb6105b68f63512703a1ed8a03cd155903" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662778 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-updater" containerID="cri-o://72b2d91a64a79c90616f09e24250e6b65a49ae69602fb773330ae5f11fcb04d4" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662807 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-auditor" containerID="cri-o://4da7895e64c6f0a051ead34f62c7e31342e26f449bd134c7e3529d04b73c2f3d" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662838 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-replicator" containerID="cri-o://6c9a46ab54e3016848961bed1d5d1cc0eeddcff8edf1e007687b96267338090d" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662879 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-server" containerID="cri-o://336ce838d7770e81da4c7557a31fa795b90be60a48b3e9b579e1d82a0a17207a" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662893 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-auditor" containerID="cri-o://62177bc6b3f5210e61d4fa109ad0edab21794e6c92a8e65133e7ae6f064ed8f9" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662948 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-replicator" containerID="cri-o://7470ca0d05ae46b498189734f3f7a811e3f808a06e71678312140fd5f7c389ea" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662964 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-server" containerID="cri-o://013814fda0f11cb6869f76d4caabd04fe6d67c24b613c2843c512e76f378e1d2" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.663029 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-replicator" containerID="cri-o://804fe60a8b7c232d84bf3cd9ed92ad47513e9d616208774db899e1a670a8fa9d" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.663077 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-auditor" containerID="cri-o://290870635c010a1c742ef38d88d7a95bc4bda2febe2172b79b9d8b8044db9b56" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.662937 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-reaper" containerID="cri-o://8f1dd12bc897fdadff2fb989c2bead87e41df04b90e728e0cf683bc33de025c4" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.681944 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a is running failed: container process not found" containerID="ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.686294 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a is running failed: container process not found" containerID="ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.698758 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a is running failed: container process not found" containerID="ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.698834 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerName="ovsdbserver-sb" Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.716287 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-bl2rc"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.795081 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-55cdcb545c-rfq2m"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.795372 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-55cdcb545c-rfq2m" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerName="neutron-api" containerID="cri-o://fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.795787 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-55cdcb545c-rfq2m" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerName="neutron-httpd" containerID="cri-o://886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97" gracePeriod=30 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.916249 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-x9khq"] Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.916613 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" podUID="4c47d914-960a-4c5e-a6f3-584998910e09" containerName="dnsmasq-dns" containerID="cri-o://e953a76ab2110b06b5538727038bb150355effbb23a98d4d39a75531a080b3b0" gracePeriod=10 Oct 04 03:25:32 crc kubenswrapper[4770]: I1004 03:25:32.976870 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhxhj\" (UniqueName: \"kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj\") pod \"novacell109d5-account-delete-6g4cl\" (UID: \"0f56d0a9-6e93-47b0-8ba2-4b04020f9296\") " pod="openstack/novacell109d5-account-delete-6g4cl" Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.984508 4770 projected.go:194] Error preparing data for projected volume kube-api-access-qhxhj for pod openstack/novacell109d5-account-delete-6g4cl: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 04 03:25:32 crc kubenswrapper[4770]: E1004 03:25:32.984589 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj podName:0f56d0a9-6e93-47b0-8ba2-4b04020f9296 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:33.984567646 +0000 UTC m=+1345.276577358 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-qhxhj" (UniqueName: "kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj") pod "novacell109d5-account-delete-6g4cl" (UID: "0f56d0a9-6e93-47b0-8ba2-4b04020f9296") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.064304 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-kgblq"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.093621 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-kgblq"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.132079 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-4rvtr"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.146742 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-4rvtr"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.149481 4770 generic.go:334] "Generic (PLEG): container finished" podID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerID="886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.149569 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55cdcb545c-rfq2m" event={"ID":"145bb94b-ff5d-4b2f-97b1-1b225fc4f164","Type":"ContainerDied","Data":"886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.152755 4770 generic.go:334] "Generic (PLEG): container finished" podID="4c47d914-960a-4c5e-a6f3-584998910e09" containerID="e953a76ab2110b06b5538727038bb150355effbb23a98d4d39a75531a080b3b0" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.152824 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" event={"ID":"4c47d914-960a-4c5e-a6f3-584998910e09","Type":"ContainerDied","Data":"e953a76ab2110b06b5538727038bb150355effbb23a98d4d39a75531a080b3b0"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.154855 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-tn765_b4444038-6604-4e51-b27d-20d3616bf640/openstack-network-exporter/0.log" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.154891 4770 generic.go:334] "Generic (PLEG): container finished" podID="b4444038-6604-4e51-b27d-20d3616bf640" containerID="3089621a8e69e00ac3e8841ec02e869f55aaa2a7703a475f407db0d736035cd6" exitCode=2 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.154953 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-tn765" event={"ID":"b4444038-6604-4e51-b27d-20d3616bf640","Type":"ContainerDied","Data":"3089621a8e69e00ac3e8841ec02e869f55aaa2a7703a475f407db0d736035cd6"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.177327 4770 generic.go:334] "Generic (PLEG): container finished" podID="5d67ab7f-c0b5-476c-b5d2-99b90eed466a" containerID="f4087f933afeeb3b260782f8334f21234028523075122f703a71f7c541bb19dc" exitCode=137 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.194925 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-576495ccf7-bslzp"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.212181 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-dbfd75b46-z42g2"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.212811 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-dbfd75b46-z42g2" podUID="76f9865e-be12-4821-a088-956061126f7b" containerName="placement-log" containerID="cri-o://bc0827aed61857d8364429abe1f13dfc75ffc958b4f218ff97327c98f09244f0" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.213368 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-dbfd75b46-z42g2" podUID="76f9865e-be12-4821-a088-956061126f7b" containerName="placement-api" containerID="cri-o://2ed2b0c7307cdff4de6c68749e0b18dddacb71011729c291fbd02e14c6b6f756" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.215934 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" containerID="cri-o://51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" gracePeriod=29 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.225387 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.225813 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerName="glance-log" containerID="cri-o://e8c9ef6e4c5a603a6c8991200968c6b52e850056eec488d0768757430caac141" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.226604 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerName="glance-httpd" containerID="cri-o://4e30affad8c19685e28c648ecf8ff6c026d5dda9129c4944f788d2e762582b04" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.241447 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.241872 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerName="glance-log" containerID="cri-o://b945952abce048f596643629d0b14f695125f7fc6eddcf6aec9901c2860f9cbb" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.242362 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerName="glance-httpd" containerID="cri-o://1f8470160c79b09e183b8b0818cd1ff1775ade9ca40792f2aa78671ca11f3a5b" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244539 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="6385e85337f2e8e8a6703c8fe588bbfb6105b68f63512703a1ed8a03cd155903" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244588 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="72b2d91a64a79c90616f09e24250e6b65a49ae69602fb773330ae5f11fcb04d4" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244598 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="4da7895e64c6f0a051ead34f62c7e31342e26f449bd134c7e3529d04b73c2f3d" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244605 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="6c9a46ab54e3016848961bed1d5d1cc0eeddcff8edf1e007687b96267338090d" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244612 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="0bbc28ed4f22995dd014ebc4247f63d2b41fd2841242ce9db1f8c81ed20dd754" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244620 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="62177bc6b3f5210e61d4fa109ad0edab21794e6c92a8e65133e7ae6f064ed8f9" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244626 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="7470ca0d05ae46b498189734f3f7a811e3f808a06e71678312140fd5f7c389ea" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244634 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="8f1dd12bc897fdadff2fb989c2bead87e41df04b90e728e0cf683bc33de025c4" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244918 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="290870635c010a1c742ef38d88d7a95bc4bda2febe2172b79b9d8b8044db9b56" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.244927 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="804fe60a8b7c232d84bf3cd9ed92ad47513e9d616208774db899e1a670a8fa9d" exitCode=0 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245085 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"6385e85337f2e8e8a6703c8fe588bbfb6105b68f63512703a1ed8a03cd155903"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245131 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"72b2d91a64a79c90616f09e24250e6b65a49ae69602fb773330ae5f11fcb04d4"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245144 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"4da7895e64c6f0a051ead34f62c7e31342e26f449bd134c7e3529d04b73c2f3d"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245153 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"6c9a46ab54e3016848961bed1d5d1cc0eeddcff8edf1e007687b96267338090d"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245166 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"0bbc28ed4f22995dd014ebc4247f63d2b41fd2841242ce9db1f8c81ed20dd754"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245174 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"62177bc6b3f5210e61d4fa109ad0edab21794e6c92a8e65133e7ae6f064ed8f9"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245184 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"7470ca0d05ae46b498189734f3f7a811e3f808a06e71678312140fd5f7c389ea"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245194 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"8f1dd12bc897fdadff2fb989c2bead87e41df04b90e728e0cf683bc33de025c4"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245204 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"290870635c010a1c742ef38d88d7a95bc4bda2febe2172b79b9d8b8044db9b56"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.245213 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"804fe60a8b7c232d84bf3cd9ed92ad47513e9d616208774db899e1a670a8fa9d"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.263990 4770 generic.go:334] "Generic (PLEG): container finished" podID="87293d1f-5e33-4a9a-9978-67487fd14809" containerID="68b1223ff961d5213ea6bc5690e64d21567c84142a78aaf580e2d2077be0f0d2" exitCode=143 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.264105 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87293d1f-5e33-4a9a-9978-67487fd14809","Type":"ContainerDied","Data":"68b1223ff961d5213ea6bc5690e64d21567c84142a78aaf580e2d2077be0f0d2"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.280540 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec07-account-create-zfvpg"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.294514 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ec07-account-create-zfvpg"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.294557 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-576495ccf7-bslzp" event={"ID":"8207573d-3535-4814-aff9-15c30c7600c7","Type":"ContainerStarted","Data":"6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.294583 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-576495ccf7-bslzp" event={"ID":"8207573d-3535-4814-aff9-15c30c7600c7","Type":"ContainerStarted","Data":"867e45b18f1a5f2f1163a002ede09e816db61b414739946c55e117f8d21a0d37"} Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.308193 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-slblm"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.319555 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderec07-account-delete-j44sn"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.332755 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-slblm"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.340572 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_57f70627-2bdd-4780-8044-75c3de4aee05/ovsdbserver-nb/0.log" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.340655 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.354676 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d4508feb-4ff8-47ec-86d7-7a9a0ad61a24/ovsdbserver-sb/0.log" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.354889 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.374661 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-v872z"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394388 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"57f70627-2bdd-4780-8044-75c3de4aee05\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394427 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394479 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdbserver-sb-tls-certs\") pod \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394525 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-metrics-certs-tls-certs\") pod \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394561 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-scripts\") pod \"57f70627-2bdd-4780-8044-75c3de4aee05\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394611 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdbserver-nb-tls-certs\") pod \"57f70627-2bdd-4780-8044-75c3de4aee05\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394642 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtpmm\" (UniqueName: \"kubernetes.io/projected/57f70627-2bdd-4780-8044-75c3de4aee05-kube-api-access-jtpmm\") pod \"57f70627-2bdd-4780-8044-75c3de4aee05\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394683 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdb-rundir\") pod \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394706 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-config\") pod \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394769 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdb-rundir\") pod \"57f70627-2bdd-4780-8044-75c3de4aee05\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394796 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxng7\" (UniqueName: \"kubernetes.io/projected/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-kube-api-access-rxng7\") pod \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394835 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-config\") pod \"57f70627-2bdd-4780-8044-75c3de4aee05\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394865 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-metrics-certs-tls-certs\") pod \"57f70627-2bdd-4780-8044-75c3de4aee05\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394891 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-combined-ca-bundle\") pod \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394958 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-scripts\") pod \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\" (UID: \"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.394985 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-combined-ca-bundle\") pod \"57f70627-2bdd-4780-8044-75c3de4aee05\" (UID: \"57f70627-2bdd-4780-8044-75c3de4aee05\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.397178 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" (UID: "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.398122 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-scripts" (OuterVolumeSpecName: "scripts") pod "57f70627-2bdd-4780-8044-75c3de4aee05" (UID: "57f70627-2bdd-4780-8044-75c3de4aee05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.399988 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-config" (OuterVolumeSpecName: "config") pod "57f70627-2bdd-4780-8044-75c3de4aee05" (UID: "57f70627-2bdd-4780-8044-75c3de4aee05"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.400065 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "57f70627-2bdd-4780-8044-75c3de4aee05" (UID: "57f70627-2bdd-4780-8044-75c3de4aee05"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.400538 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-config" (OuterVolumeSpecName: "config") pod "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" (UID: "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.401297 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "57f70627-2bdd-4780-8044-75c3de4aee05" (UID: "57f70627-2bdd-4780-8044-75c3de4aee05"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.402044 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-scripts" (OuterVolumeSpecName: "scripts") pod "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" (UID: "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.416511 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57f70627-2bdd-4780-8044-75c3de4aee05-kube-api-access-jtpmm" (OuterVolumeSpecName: "kube-api-access-jtpmm") pod "57f70627-2bdd-4780-8044-75c3de4aee05" (UID: "57f70627-2bdd-4780-8044-75c3de4aee05"). InnerVolumeSpecName "kube-api-access-jtpmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.417130 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-kube-api-access-rxng7" (OuterVolumeSpecName: "kube-api-access-rxng7") pod "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" (UID: "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24"). InnerVolumeSpecName "kube-api-access-rxng7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.422361 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-v872z"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.425517 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-tn765_b4444038-6604-4e51-b27d-20d3616bf640/openstack-network-exporter/0.log" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.425668 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.425513 4770 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 04 03:25:33 crc kubenswrapper[4770]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 04 03:25:33 crc kubenswrapper[4770]: + source /usr/local/bin/container-scripts/functions Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNBridge=br-int Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNRemote=tcp:localhost:6642 Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNEncapType=geneve Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNAvailabilityZones= Oct 04 03:25:33 crc kubenswrapper[4770]: ++ EnableChassisAsGateway=true Oct 04 03:25:33 crc kubenswrapper[4770]: ++ PhysicalNetworks= Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNHostName= Oct 04 03:25:33 crc kubenswrapper[4770]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 04 03:25:33 crc kubenswrapper[4770]: ++ ovs_dir=/var/lib/openvswitch Oct 04 03:25:33 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 04 03:25:33 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 04 03:25:33 crc kubenswrapper[4770]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + sleep 0.5 Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + sleep 0.5 Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + sleep 0.5 Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + cleanup_ovsdb_server_semaphore Oct 04 03:25:33 crc kubenswrapper[4770]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 04 03:25:33 crc kubenswrapper[4770]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 04 03:25:33 crc kubenswrapper[4770]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-6nz6q" message=< Oct 04 03:25:33 crc kubenswrapper[4770]: Exiting ovsdb-server (5) [ OK ] Oct 04 03:25:33 crc kubenswrapper[4770]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 04 03:25:33 crc kubenswrapper[4770]: + source /usr/local/bin/container-scripts/functions Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNBridge=br-int Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNRemote=tcp:localhost:6642 Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNEncapType=geneve Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNAvailabilityZones= Oct 04 03:25:33 crc kubenswrapper[4770]: ++ EnableChassisAsGateway=true Oct 04 03:25:33 crc kubenswrapper[4770]: ++ PhysicalNetworks= Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNHostName= Oct 04 03:25:33 crc kubenswrapper[4770]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 04 03:25:33 crc kubenswrapper[4770]: ++ ovs_dir=/var/lib/openvswitch Oct 04 03:25:33 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 04 03:25:33 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 04 03:25:33 crc kubenswrapper[4770]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + sleep 0.5 Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + sleep 0.5 Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + sleep 0.5 Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + cleanup_ovsdb_server_semaphore Oct 04 03:25:33 crc kubenswrapper[4770]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 04 03:25:33 crc kubenswrapper[4770]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 04 03:25:33 crc kubenswrapper[4770]: > Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.427443 4770 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 04 03:25:33 crc kubenswrapper[4770]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 04 03:25:33 crc kubenswrapper[4770]: + source /usr/local/bin/container-scripts/functions Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNBridge=br-int Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNRemote=tcp:localhost:6642 Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNEncapType=geneve Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNAvailabilityZones= Oct 04 03:25:33 crc kubenswrapper[4770]: ++ EnableChassisAsGateway=true Oct 04 03:25:33 crc kubenswrapper[4770]: ++ PhysicalNetworks= Oct 04 03:25:33 crc kubenswrapper[4770]: ++ OVNHostName= Oct 04 03:25:33 crc kubenswrapper[4770]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 04 03:25:33 crc kubenswrapper[4770]: ++ ovs_dir=/var/lib/openvswitch Oct 04 03:25:33 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 04 03:25:33 crc kubenswrapper[4770]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 04 03:25:33 crc kubenswrapper[4770]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + sleep 0.5 Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + sleep 0.5 Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + sleep 0.5 Oct 04 03:25:33 crc kubenswrapper[4770]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 04 03:25:33 crc kubenswrapper[4770]: + cleanup_ovsdb_server_semaphore Oct 04 03:25:33 crc kubenswrapper[4770]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 04 03:25:33 crc kubenswrapper[4770]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 04 03:25:33 crc kubenswrapper[4770]: > pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" containerID="cri-o://4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.428435 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" containerID="cri-o://4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" gracePeriod=29 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.440999 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5365-account-create-47djg"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.447339 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" (UID: "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.472567 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5365-account-create-47djg"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.495167 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.512579 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-metrics-certs-tls-certs\") pod \"b4444038-6604-4e51-b27d-20d3616bf640\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.513331 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-combined-ca-bundle\") pod \"b4444038-6604-4e51-b27d-20d3616bf640\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.513514 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovs-rundir\") pod \"b4444038-6604-4e51-b27d-20d3616bf640\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.513646 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovn-rundir\") pod \"b4444038-6604-4e51-b27d-20d3616bf640\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.513686 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh8bv\" (UniqueName: \"kubernetes.io/projected/b4444038-6604-4e51-b27d-20d3616bf640-kube-api-access-kh8bv\") pod \"b4444038-6604-4e51-b27d-20d3616bf640\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.513730 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4444038-6604-4e51-b27d-20d3616bf640-config\") pod \"b4444038-6604-4e51-b27d-20d3616bf640\" (UID: \"b4444038-6604-4e51-b27d-20d3616bf640\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.514091 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "b4444038-6604-4e51-b27d-20d3616bf640" (UID: "b4444038-6604-4e51-b27d-20d3616bf640"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.514141 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "b4444038-6604-4e51-b27d-20d3616bf640" (UID: "b4444038-6604-4e51-b27d-20d3616bf640"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515043 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxng7\" (UniqueName: \"kubernetes.io/projected/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-kube-api-access-rxng7\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515071 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515085 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515096 4770 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515122 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515141 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515153 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b4444038-6604-4e51-b27d-20d3616bf640-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515165 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57f70627-2bdd-4780-8044-75c3de4aee05-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515176 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtpmm\" (UniqueName: \"kubernetes.io/projected/57f70627-2bdd-4780-8044-75c3de4aee05-kube-api-access-jtpmm\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515187 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515198 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.515211 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.516449 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4444038-6604-4e51-b27d-20d3616bf640-config" (OuterVolumeSpecName: "config") pod "b4444038-6604-4e51-b27d-20d3616bf640" (UID: "b4444038-6604-4e51-b27d-20d3616bf640"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.523723 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-787bf4d777-x8gln"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.524040 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-787bf4d777-x8gln" podUID="06da071b-e1a1-42da-9e6e-fd957923876d" containerName="proxy-httpd" containerID="cri-o://8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.524443 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-787bf4d777-x8gln" podUID="06da071b-e1a1-42da-9e6e-fd957923876d" containerName="proxy-server" containerID="cri-o://5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.562741 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b0c9-account-create-dx2vq"] Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.567504 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.585182 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.586312 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-55cdcb545c-rfq2m" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.165:9696/\": dial tcp 10.217.0.165:9696: connect: connection refused" Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.586325 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.586384 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.586449 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.592398 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.597916 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4444038-6604-4e51-b27d-20d3616bf640-kube-api-access-kh8bv" (OuterVolumeSpecName: "kube-api-access-kh8bv") pod "b4444038-6604-4e51-b27d-20d3616bf640" (UID: "b4444038-6604-4e51-b27d-20d3616bf640"). InnerVolumeSpecName "kube-api-access-kh8bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.599775 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b0c9-account-create-dx2vq"] Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.608347 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.608444 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.619140 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh8bv\" (UniqueName: \"kubernetes.io/projected/b4444038-6604-4e51-b27d-20d3616bf640-kube-api-access-kh8bv\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.619173 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4444038-6604-4e51-b27d-20d3616bf640-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.628041 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57f70627-2bdd-4780-8044-75c3de4aee05" (UID: "57f70627-2bdd-4780-8044-75c3de4aee05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.637159 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronb0c9-account-delete-hc97s"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.650545 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-dtz9z"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.662206 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-dtz9z"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.693991 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32226404-ae26-443d-aec7-766409996096" path="/var/lib/kubelet/pods/32226404-ae26-443d-aec7-766409996096/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.695238 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47a8258e-5ecc-44eb-8c4d-20ecd8078cb9" path="/var/lib/kubelet/pods/47a8258e-5ecc-44eb-8c4d-20ecd8078cb9/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.696933 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="483a5eae-40e5-4e11-82e8-24198ec8a733" path="/var/lib/kubelet/pods/483a5eae-40e5-4e11-82e8-24198ec8a733/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.697624 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6df7b787-d263-4732-a672-66a0b33c6369" path="/var/lib/kubelet/pods/6df7b787-d263-4732-a672-66a0b33c6369/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.703264 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76095968-a5a2-4a39-bf86-5faafb9d76d3" path="/var/lib/kubelet/pods/76095968-a5a2-4a39-bf86-5faafb9d76d3/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.704294 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c0eb670-7851-4fdc-92ce-d32bcb5e79e9" path="/var/lib/kubelet/pods/7c0eb670-7851-4fdc-92ce-d32bcb5e79e9/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.705382 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8063ce88-486b-413b-b13f-ac3c855fce17" path="/var/lib/kubelet/pods/8063ce88-486b-413b-b13f-ac3c855fce17/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.706098 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8acc0024-f6e4-44a4-a76d-007c7ff3ecd8" path="/var/lib/kubelet/pods/8acc0024-f6e4-44a4-a76d-007c7ff3ecd8/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.710059 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c51889c-6112-4c86-93f1-a69876016f4f" path="/var/lib/kubelet/pods/9c51889c-6112-4c86-93f1-a69876016f4f/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.711500 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0443425-c134-49da-963c-79c4188e9780" path="/var/lib/kubelet/pods/a0443425-c134-49da-963c-79c4188e9780/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.723248 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6129f31-e510-495e-a4ad-2976f2054635" path="/var/lib/kubelet/pods/e6129f31-e510-495e-a4ad-2976f2054635/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.725654 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:33 crc kubenswrapper[4770]: E1004 03:25:33.725739 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data podName:642fe06d-e5b8-461e-a1f7-dbcefc071945 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:35.725721522 +0000 UTC m=+1347.017731234 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data") pod "rabbitmq-cell1-server-0" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945") : configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.726082 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.729279 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec2a2801-f563-4bbf-b191-7e4dffbe89a2" path="/var/lib/kubelet/pods/ec2a2801-f563-4bbf-b191-7e4dffbe89a2/volumes" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.750673 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementded7-account-delete-qxzt9"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.750715 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-bgpjr"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.750728 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-bgpjr"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.750741 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ded7-account-create-cbnfn"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.750752 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ded7-account-create-cbnfn"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.750762 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-wnwk5"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.750772 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-wnwk5"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.750782 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-d8ca-account-create-kj8sq"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.763152 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.765458 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-d8ca-account-create-kj8sq"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.781787 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glanced8ca-account-delete-k2b6r"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.787734 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.796855 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.797317 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-log" containerID="cri-o://30a450deeec6b43046f1e4c12cadb70ffef33664a48ce36b69bf552dadbd53c4" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.797705 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-api" containerID="cri-o://6bba8b83577d71e684e12c6e57b942c931915975b422c9aa9892702880a260fc" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.827815 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppvfm\" (UniqueName: \"kubernetes.io/projected/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-kube-api-access-ppvfm\") pod \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.827864 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config\") pod \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.828076 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-combined-ca-bundle\") pod \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.828169 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config-secret\") pod \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\" (UID: \"5d67ab7f-c0b5-476c-b5d2-99b90eed466a\") " Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.840153 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-pv8zj"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.876789 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-kube-api-access-ppvfm" (OuterVolumeSpecName: "kube-api-access-ppvfm") pod "5d67ab7f-c0b5-476c-b5d2-99b90eed466a" (UID: "5d67ab7f-c0b5-476c-b5d2-99b90eed466a"). InnerVolumeSpecName "kube-api-access-ppvfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.886964 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-pv8zj"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.919656 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.920128 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-log" containerID="cri-o://1e4a44749d34e20093fa0ef2448d2cdc6233bb16e1aa47b08bdfb010277a16ae" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.921462 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-metadata" containerID="cri-o://fa7c4e0199de7a74a6a51a933f598585e135eac2ee7b99b1d3d49fa44c55ad5a" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.935711 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppvfm\" (UniqueName: \"kubernetes.io/projected/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-kube-api-access-ppvfm\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.952896 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-a2a5-account-create-d6vtx"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.979339 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-a2a5-account-create-d6vtx"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.979404 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f85c64c7b-jcqvq"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.979617 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f85c64c7b-jcqvq" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api-log" containerID="cri-o://76df273164d25e12be0bd5fba204411c763fba6be089bd2cf529d36352a6edac" gracePeriod=30 Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.984643 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-09d5-account-create-5kncn"] Oct 04 03:25:33 crc kubenswrapper[4770]: I1004 03:25:33.984683 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f85c64c7b-jcqvq" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api" containerID="cri-o://f652f94cc4c1e2df6ea6011817891a4c91174bbbbee764ff5579bda21b7ab8fb" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.007327 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" (UID: "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.037485 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhxhj\" (UniqueName: \"kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj\") pod \"novacell109d5-account-delete-6g4cl\" (UID: \"0f56d0a9-6e93-47b0-8ba2-4b04020f9296\") " pod="openstack/novacell109d5-account-delete-6g4cl" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.038141 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.040919 4770 projected.go:194] Error preparing data for projected volume kube-api-access-qhxhj for pod openstack/novacell109d5-account-delete-6g4cl: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.040969 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj podName:0f56d0a9-6e93-47b0-8ba2-4b04020f9296 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:36.040952739 +0000 UTC m=+1347.332962451 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-qhxhj" (UniqueName: "kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj") pod "novacell109d5-account-delete-6g4cl" (UID: "0f56d0a9-6e93-47b0-8ba2-4b04020f9296") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.042197 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="2def82a8-5927-4ebe-ac87-e8ad232797ee" containerName="rabbitmq" containerID="cri-o://19babfd2683be969879e3923b984d72aa3faecdef7166dfa931d52c111a1f74f" gracePeriod=604800 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.070423 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.108102 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-09d5-account-create-5kncn"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.127167 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.127543 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="41a2c8f7-f9d1-44bf-8381-ceea9504a909" containerName="nova-scheduler-scheduler" containerID="cri-o://c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.140182 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.148893 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5d67ab7f-c0b5-476c-b5d2-99b90eed466a" (UID: "5d67ab7f-c0b5-476c-b5d2-99b90eed466a"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.149160 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d67ab7f-c0b5-476c-b5d2-99b90eed466a" (UID: "5d67ab7f-c0b5-476c-b5d2-99b90eed466a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.158376 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4444038-6604-4e51-b27d-20d3616bf640" (UID: "b4444038-6604-4e51-b27d-20d3616bf640"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.162319 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" (UID: "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.172810 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.183892 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-kv9tv"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.194038 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell109d5-account-delete-6g4cl"] Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.195206 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-qhxhj], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/novacell109d5-account-delete-6g4cl" podUID="0f56d0a9-6e93-47b0-8ba2-4b04020f9296" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.200326 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-kv9tv"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.203090 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="2def82a8-5927-4ebe-ac87-e8ad232797ee" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.218709 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-66b9dbf78f-gv52v"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.218963 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-66b9dbf78f-gv52v" podUID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerName="barbican-worker-log" containerID="cri-o://ad1c6f5e9093316bec8bf19c97c384353c2014b68fb11557c892fae059e000b7" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.219314 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-66b9dbf78f-gv52v" podUID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerName="barbican-worker" containerID="cri-o://ca61c9dff3e5c232108ba39a8b1a8f4f4f5b0eb23bc5838c1ee8c7e06106319e" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.233166 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5d67ab7f-c0b5-476c-b5d2-99b90eed466a" (UID: "5d67ab7f-c0b5-476c-b5d2-99b90eed466a"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: W1004 03:25:34.239190 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc00a8476_9758_46f5_864f_9a7b80ca8bd1.slice/crio-0059ba19a9d39f65a573385056c75a8ed13f8e1072497933e7398653b1dac84e WatchSource:0}: Error finding container 0059ba19a9d39f65a573385056c75a8ed13f8e1072497933e7398653b1dac84e: Status 404 returned error can't find the container with id 0059ba19a9d39f65a573385056c75a8ed13f8e1072497933e7398653b1dac84e Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.242103 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-576495ccf7-bslzp"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.243104 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.243143 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.243156 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.243169 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.243186 4770 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.243199 4770 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d67ab7f-c0b5-476c-b5d2-99b90eed466a-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.268780 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5877f57886-5xk9s"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.269688 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" podUID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerName="barbican-keystone-listener-log" containerID="cri-o://3f304c4c9d4084df4db4c9d1ca2bc9d44285df6bb49699c82b5e8823e8e1ae3f" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.272029 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" podUID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerName="barbican-keystone-listener" containerID="cri-o://ba22ef47e0582d9695bafcc9660a1b4adb815a65c5ef9b6fd33f435fbfb67a12" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.276146 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-867bd5444b-xx8d4"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.282913 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.324206 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.324658 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="1d236ba3-55db-460a-9daa-d6e6e8b06766" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.364313 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-576495ccf7-bslzp" event={"ID":"8207573d-3535-4814-aff9-15c30c7600c7","Type":"ContainerStarted","Data":"897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.365786 4770 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican-worker-576495ccf7-bslzp" secret="" err="secret \"barbican-barbican-dockercfg-8l96n\" not found" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.368530 4770 generic.go:334] "Generic (PLEG): container finished" podID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerID="e8c9ef6e4c5a603a6c8991200968c6b52e850056eec488d0768757430caac141" exitCode=143 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.368595 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d49edb80-7b9b-4b10-afbc-652ec540542e","Type":"ContainerDied","Data":"e8c9ef6e4c5a603a6c8991200968c6b52e850056eec488d0768757430caac141"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.380452 4770 generic.go:334] "Generic (PLEG): container finished" podID="76f9865e-be12-4821-a088-956061126f7b" containerID="bc0827aed61857d8364429abe1f13dfc75ffc958b4f218ff97327c98f09244f0" exitCode=143 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.380501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dbfd75b46-z42g2" event={"ID":"76f9865e-be12-4821-a088-956061126f7b","Type":"ContainerDied","Data":"bc0827aed61857d8364429abe1f13dfc75ffc958b4f218ff97327c98f09244f0"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.392564 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.392815 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="ef58a910-f85e-4e14-95b5-807741097485" containerName="nova-cell1-conductor-conductor" containerID="cri-o://7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.402116 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3321-account-delete-gmw9l" event={"ID":"6d5328f9-e5e9-4ad5-96ff-02442117ff0e","Type":"ContainerStarted","Data":"e3c582c0354b608d8eac07ebe1b7bb0a1cf369dc801f51dac96dcff1d7f893d3"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.404068 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcntx"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.410943 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="642fe06d-e5b8-461e-a1f7-dbcefc071945" containerName="rabbitmq" containerID="cri-o://4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749" gracePeriod=604800 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.413745 4770 generic.go:334] "Generic (PLEG): container finished" podID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerID="1e4a44749d34e20093fa0ef2448d2cdc6233bb16e1aa47b08bdfb010277a16ae" exitCode=143 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.413832 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79ab7038-39be-4c09-82ca-00f83aa69e37","Type":"ContainerDied","Data":"1e4a44749d34e20093fa0ef2448d2cdc6233bb16e1aa47b08bdfb010277a16ae"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.421666 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="35f43e87-e3d9-4508-8bb1-4fe89c64017f" containerName="galera" containerID="cri-o://0218e06d6c52fad79e58c3f1c9b92b86ec1c9c65d56610da75164f0326b481a0" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.428244 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rcntx"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.430906 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" event={"ID":"4c47d914-960a-4c5e-a6f3-584998910e09","Type":"ContainerDied","Data":"0a9006b887e7f4386afae8b76b8ef40d5a366866c54ce68deaebb74622b739fc"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.430930 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "57f70627-2bdd-4780-8044-75c3de4aee05" (UID: "57f70627-2bdd-4780-8044-75c3de4aee05"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.430963 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a9006b887e7f4386afae8b76b8ef40d5a366866c54ce68deaebb74622b739fc" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.433385 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "57f70627-2bdd-4780-8044-75c3de4aee05" (UID: "57f70627-2bdd-4780-8044-75c3de4aee05"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.436358 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" (UID: "d4508feb-4ff8-47ec-86d7-7a9a0ad61a24"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.447232 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pz6zp"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.454629 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.454669 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/57f70627-2bdd-4780-8044-75c3de4aee05-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.454682 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.458415 4770 generic.go:334] "Generic (PLEG): container finished" podID="06da071b-e1a1-42da-9e6e-fd957923876d" containerID="8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70" exitCode=0 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.458559 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-787bf4d777-x8gln" event={"ID":"06da071b-e1a1-42da-9e6e-fd957923876d","Type":"ContainerDied","Data":"8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.467090 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.467629 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="e826d9bb-183e-401b-bc8b-6cfe0938a7f6" containerName="nova-cell0-conductor-conductor" containerID="cri-o://aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092" gracePeriod=30 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.480781 4770 generic.go:334] "Generic (PLEG): container finished" podID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerID="76df273164d25e12be0bd5fba204411c763fba6be089bd2cf529d36352a6edac" exitCode=143 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.480858 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f85c64c7b-jcqvq" event={"ID":"83d745cd-57f5-4624-9d2c-75353e52fa56","Type":"ContainerDied","Data":"76df273164d25e12be0bd5fba204411c763fba6be089bd2cf529d36352a6edac"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.484353 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pz6zp"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.486691 4770 generic.go:334] "Generic (PLEG): container finished" podID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerID="b945952abce048f596643629d0b14f695125f7fc6eddcf6aec9901c2860f9cbb" exitCode=143 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.486857 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d","Type":"ContainerDied","Data":"b945952abce048f596643629d0b14f695125f7fc6eddcf6aec9901c2860f9cbb"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.510279 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderec07-account-delete-j44sn" event={"ID":"2af9166c-e2c5-462d-be52-340be8aadfd5","Type":"ContainerStarted","Data":"3409dbb112e2de95127ddfe61612a916b1c335555b22af005a35cb7cfe869fce"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.534658 4770 generic.go:334] "Generic (PLEG): container finished" podID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" exitCode=0 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.534754 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6nz6q" event={"ID":"c5f2e918-133c-440e-ab52-2a389ddc9ede","Type":"ContainerDied","Data":"4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.537921 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderec07-account-delete-j44sn"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.546073 4770 scope.go:117] "RemoveContainer" containerID="f4087f933afeeb3b260782f8334f21234028523075122f703a71f7c541bb19dc" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.546241 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.553114 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "b4444038-6604-4e51-b27d-20d3616bf640" (UID: "b4444038-6604-4e51-b27d-20d3616bf640"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.557907 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4444038-6604-4e51-b27d-20d3616bf640-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.563498 4770 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.563569 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:35.063549868 +0000 UTC m=+1346.355559580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-worker-config-data" not found Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.563825 4770 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.563858 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:35.063849806 +0000 UTC m=+1346.355859518 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-config-data" not found Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.572118 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronb0c9-account-delete-hc97s"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.580793 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" event={"ID":"9b1ac142-741d-48a1-bf5c-982bbf7bace6","Type":"ContainerStarted","Data":"d6a49ed7da5a07ce7333bc8b047f047c4f024bac87ba65085c3f9244b5c2353f"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.583480 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glanced8ca-account-delete-k2b6r"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.591107 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi3321-account-delete-gmw9l"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.607178 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementded7-account-delete-qxzt9"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.627347 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="2aea2997649b938d2836a8f46cd25d7caab6d10e8f2aeeac5509198aa88b89ca" exitCode=0 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.627372 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="336ce838d7770e81da4c7557a31fa795b90be60a48b3e9b579e1d82a0a17207a" exitCode=0 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.627380 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="013814fda0f11cb6869f76d4caabd04fe6d67c24b613c2843c512e76f378e1d2" exitCode=0 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.627387 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="f824e75e4b4bc8b5e89c5df2fd7e73c49c2760133924bc91336e17d29f37ffa6" exitCode=0 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.627430 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"2aea2997649b938d2836a8f46cd25d7caab6d10e8f2aeeac5509198aa88b89ca"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.627462 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"336ce838d7770e81da4c7557a31fa795b90be60a48b3e9b579e1d82a0a17207a"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.627474 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"013814fda0f11cb6869f76d4caabd04fe6d67c24b613c2843c512e76f378e1d2"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.627482 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"f824e75e4b4bc8b5e89c5df2fd7e73c49c2760133924bc91336e17d29f37ffa6"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.678639 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-867bd5444b-xx8d4"] Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.683916 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.683979 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data podName:2def82a8-5927-4ebe-ac87-e8ad232797ee nodeName:}" failed. No retries permitted until 2025-10-04 03:25:38.683957047 +0000 UTC m=+1349.975966759 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data") pod "rabbitmq-server-0" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee") : configmap "rabbitmq-config-data" not found Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.699397 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-tn765_b4444038-6604-4e51-b27d-20d3616bf640/openstack-network-exporter/0.log" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.699501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-tn765" event={"ID":"b4444038-6604-4e51-b27d-20d3616bf640","Type":"ContainerDied","Data":"77c3c22bb2cdb8dbe4b9be8b7e236168e541c168357acc474291a61c021a0b13"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.699680 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-tn765" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.705618 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.722480 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-576495ccf7-bslzp" podStartSLOduration=4.722439644 podStartE2EDuration="4.722439644s" podCreationTimestamp="2025-10-04 03:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:25:34.3935147 +0000 UTC m=+1345.685524412" watchObservedRunningTime="2025-10-04 03:25:34.722439644 +0000 UTC m=+1346.014449356" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.733256 4770 generic.go:334] "Generic (PLEG): container finished" podID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerID="1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522" exitCode=0 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.733377 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2","Type":"ContainerDied","Data":"1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.742029 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementded7-account-delete-qxzt9" event={"ID":"d18c6351-6663-4a3c-84b6-2752f5e66e49","Type":"ContainerStarted","Data":"5dddefeae5fd60c8df075cb5b7feee573fd182915531c9aad3bc15288fde970d"} Oct 04 03:25:34 crc kubenswrapper[4770]: E1004 03:25:34.744091 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06da071b_e1a1_42da_9e6e_fd957923876d.slice/crio-conmon-5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode328a4dc_7a4d_4cb0_82a8_f60df78328d8.slice/crio-3f304c4c9d4084df4db4c9d1ca2bc9d44285df6bb49699c82b5e8823e8e1ae3f.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.757726 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanced8ca-account-delete-k2b6r" event={"ID":"d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c","Type":"ContainerStarted","Data":"10a76a22d6db2361ded091227c2fec250e401a12fcc957558c9406bf29541919"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.800546 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_57f70627-2bdd-4780-8044-75c3de4aee05/ovsdbserver-nb/0.log" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.800694 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"57f70627-2bdd-4780-8044-75c3de4aee05","Type":"ContainerDied","Data":"1cb22383ea5a902e5e9d36f42af1e33ae0907496e90bda94cd7c9317731febe7"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.800853 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.814780 4770 scope.go:117] "RemoveContainer" containerID="3089621a8e69e00ac3e8841ec02e869f55aaa2a7703a475f407db0d736035cd6" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.837205 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d4508feb-4ff8-47ec-86d7-7a9a0ad61a24/ovsdbserver-sb/0.log" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.837345 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d4508feb-4ff8-47ec-86d7-7a9a0ad61a24","Type":"ContainerDied","Data":"287d87a84cf1bd29f97b9071e57d29a0be857b26458c07be1832624e83dcc69b"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.837495 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.851675 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronb0c9-account-delete-hc97s" event={"ID":"c00a8476-9758-46f5-864f-9a7b80ca8bd1","Type":"ContainerStarted","Data":"0059ba19a9d39f65a573385056c75a8ed13f8e1072497933e7398653b1dac84e"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.859571 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-tn765"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.876827 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="642fe06d-e5b8-461e-a1f7-dbcefc071945" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.103:5671: connect: connection refused" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.877131 4770 generic.go:334] "Generic (PLEG): container finished" podID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerID="30a450deeec6b43046f1e4c12cadb70ffef33664a48ce36b69bf552dadbd53c4" exitCode=143 Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.877174 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b795408-0144-4d73-8a56-043ce3e38ee6","Type":"ContainerDied","Data":"30a450deeec6b43046f1e4c12cadb70ffef33664a48ce36b69bf552dadbd53c4"} Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.877211 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell109d5-account-delete-6g4cl" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.890109 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-tn765"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.900152 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell109d5-account-delete-6g4cl" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.901564 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-config\") pod \"4c47d914-960a-4c5e-a6f3-584998910e09\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.901603 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-nb\") pod \"4c47d914-960a-4c5e-a6f3-584998910e09\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.901629 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-sb\") pod \"4c47d914-960a-4c5e-a6f3-584998910e09\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.901704 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-svc\") pod \"4c47d914-960a-4c5e-a6f3-584998910e09\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.901726 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-swift-storage-0\") pod \"4c47d914-960a-4c5e-a6f3-584998910e09\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.901745 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4dpx\" (UniqueName: \"kubernetes.io/projected/4c47d914-960a-4c5e-a6f3-584998910e09-kube-api-access-p4dpx\") pod \"4c47d914-960a-4c5e-a6f3-584998910e09\" (UID: \"4c47d914-960a-4c5e-a6f3-584998910e09\") " Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.924373 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.942971 4770 scope.go:117] "RemoveContainer" containerID="760e8cc98a032092f27104dafba09f778ecc47ca79026332349329b345264e45" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.949227 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.953048 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c47d914-960a-4c5e-a6f3-584998910e09-kube-api-access-p4dpx" (OuterVolumeSpecName: "kube-api-access-p4dpx") pod "4c47d914-960a-4c5e-a6f3-584998910e09" (UID: "4c47d914-960a-4c5e-a6f3-584998910e09"). InnerVolumeSpecName "kube-api-access-p4dpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4770]: I1004 03:25:34.956790 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.000582 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.009120 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4dpx\" (UniqueName: \"kubernetes.io/projected/4c47d914-960a-4c5e-a6f3-584998910e09-kube-api-access-p4dpx\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.117523 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.139028 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.142505 4770 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.155280 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:36.155244504 +0000 UTC m=+1347.447254206 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-config-data" not found Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.144566 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.155377 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="41a2c8f7-f9d1-44bf-8381-ceea9504a909" containerName="nova-scheduler-scheduler" Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.142625 4770 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.155510 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:36.155485611 +0000 UTC m=+1347.447495323 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-worker-config-data" not found Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.151415 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4c47d914-960a-4c5e-a6f3-584998910e09" (UID: "4c47d914-960a-4c5e-a6f3-584998910e09"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.156563 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4c47d914-960a-4c5e-a6f3-584998910e09" (UID: "4c47d914-960a-4c5e-a6f3-584998910e09"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.158840 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-config" (OuterVolumeSpecName: "config") pod "4c47d914-960a-4c5e-a6f3-584998910e09" (UID: "4c47d914-960a-4c5e-a6f3-584998910e09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.170443 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c47d914-960a-4c5e-a6f3-584998910e09" (UID: "4c47d914-960a-4c5e-a6f3-584998910e09"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.206125 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4c47d914-960a-4c5e-a6f3-584998910e09" (UID: "4c47d914-960a-4c5e-a6f3-584998910e09"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.245093 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.245137 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.245182 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.245193 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.245203 4770 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c47d914-960a-4c5e-a6f3-584998910e09-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.722069 4770 scope.go:117] "RemoveContainer" containerID="92885b43eda03ead2301755a74384e53d3b27a5aa4fddd4305ff2576a9f59a7c" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.726573 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23bf2c46-426f-4bb7-82e0-f9e90d0a2f32" path="/var/lib/kubelet/pods/23bf2c46-426f-4bb7-82e0-f9e90d0a2f32/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.727083 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4da6c58e-fb44-4acb-b49d-d29a0f82882f" path="/var/lib/kubelet/pods/4da6c58e-fb44-4acb-b49d-d29a0f82882f/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.727673 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57f70627-2bdd-4780-8044-75c3de4aee05" path="/var/lib/kubelet/pods/57f70627-2bdd-4780-8044-75c3de4aee05/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.729912 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d67ab7f-c0b5-476c-b5d2-99b90eed466a" path="/var/lib/kubelet/pods/5d67ab7f-c0b5-476c-b5d2-99b90eed466a/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.730436 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6159c490-e51a-4a0b-a570-35f9bd8b82c8" path="/var/lib/kubelet/pods/6159c490-e51a-4a0b-a570-35f9bd8b82c8/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.730916 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="666786ce-354d-4d76-bf21-80d3245b7bec" path="/var/lib/kubelet/pods/666786ce-354d-4d76-bf21-80d3245b7bec/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.736725 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="670c7093-0b8d-45d2-91f1-fccd2f275adf" path="/var/lib/kubelet/pods/670c7093-0b8d-45d2-91f1-fccd2f275adf/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.740986 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1deee0b-c346-457f-94e1-7480502d0583" path="/var/lib/kubelet/pods/a1deee0b-c346-457f-94e1-7480502d0583/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.741681 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8" path="/var/lib/kubelet/pods/acc1eb6c-ed84-4a06-a8da-c8c026f5c2a8/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.745832 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b417baa3-f6c2-499f-af94-0bd91ff07e61" path="/var/lib/kubelet/pods/b417baa3-f6c2-499f-af94-0bd91ff07e61/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.750402 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4444038-6604-4e51-b27d-20d3616bf640" path="/var/lib/kubelet/pods/b4444038-6604-4e51-b27d-20d3616bf640/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.751106 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbd7579a-ef76-48fc-96f6-3b26466084f8" path="/var/lib/kubelet/pods/cbd7579a-ef76-48fc-96f6-3b26466084f8/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.751705 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" path="/var/lib/kubelet/pods/d4508feb-4ff8-47ec-86d7-7a9a0ad61a24/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.754505 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d56b3add-faa1-4600-bdca-f4bf03a11250" path="/var/lib/kubelet/pods/d56b3add-faa1-4600-bdca-f4bf03a11250/volumes" Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.756688 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:35 crc kubenswrapper[4770]: E1004 03:25:35.756748 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data podName:642fe06d-e5b8-461e-a1f7-dbcefc071945 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:39.756732898 +0000 UTC m=+1351.048742610 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data") pod "rabbitmq-cell1-server-0" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945") : configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.887667 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.161:8776/healthcheck\": read tcp 10.217.0.2:36750->10.217.0.161:8776: read: connection reset by peer" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.895235 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.895478 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.905666 4770 scope.go:117] "RemoveContainer" containerID="347a7235cf642c5bee0dc831171a48e6731de85a58d10232d67b2b94de43b821" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.941388 4770 generic.go:334] "Generic (PLEG): container finished" podID="06da071b-e1a1-42da-9e6e-fd957923876d" containerID="5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7" exitCode=0 Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.941750 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-787bf4d777-x8gln" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.941638 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-787bf4d777-x8gln" event={"ID":"06da071b-e1a1-42da-9e6e-fd957923876d","Type":"ContainerDied","Data":"5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7"} Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.942089 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-787bf4d777-x8gln" event={"ID":"06da071b-e1a1-42da-9e6e-fd957923876d","Type":"ContainerDied","Data":"38c0143552fdfa921f35db23330f1117d2705fefd42165f2e2c2ee0e96be9c27"} Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.946987 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" event={"ID":"9b1ac142-741d-48a1-bf5c-982bbf7bace6","Type":"ContainerStarted","Data":"da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97"} Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.960369 4770 generic.go:334] "Generic (PLEG): container finished" podID="d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c" containerID="4435c5f48a844ce2bd51729a6c5ff1242e1229d3621dec9c5c639c58455cb623" exitCode=0 Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.960459 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanced8ca-account-delete-k2b6r" event={"ID":"d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c","Type":"ContainerDied","Data":"4435c5f48a844ce2bd51729a6c5ff1242e1229d3621dec9c5c639c58455cb623"} Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961396 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-internal-tls-certs\") pod \"06da071b-e1a1-42da-9e6e-fd957923876d\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961457 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-run-httpd\") pod \"06da071b-e1a1-42da-9e6e-fd957923876d\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961489 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phl9r\" (UniqueName: \"kubernetes.io/projected/1d236ba3-55db-460a-9daa-d6e6e8b06766-kube-api-access-phl9r\") pod \"1d236ba3-55db-460a-9daa-d6e6e8b06766\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961550 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-etc-swift\") pod \"06da071b-e1a1-42da-9e6e-fd957923876d\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961613 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-nova-novncproxy-tls-certs\") pod \"1d236ba3-55db-460a-9daa-d6e6e8b06766\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961715 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-log-httpd\") pod \"06da071b-e1a1-42da-9e6e-fd957923876d\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961748 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-config-data\") pod \"1d236ba3-55db-460a-9daa-d6e6e8b06766\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961774 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx8hd\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-kube-api-access-rx8hd\") pod \"06da071b-e1a1-42da-9e6e-fd957923876d\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961798 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-config-data\") pod \"06da071b-e1a1-42da-9e6e-fd957923876d\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961820 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-public-tls-certs\") pod \"06da071b-e1a1-42da-9e6e-fd957923876d\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961846 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-combined-ca-bundle\") pod \"1d236ba3-55db-460a-9daa-d6e6e8b06766\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961869 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-combined-ca-bundle\") pod \"06da071b-e1a1-42da-9e6e-fd957923876d\" (UID: \"06da071b-e1a1-42da-9e6e-fd957923876d\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.961947 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-vencrypt-tls-certs\") pod \"1d236ba3-55db-460a-9daa-d6e6e8b06766\" (UID: \"1d236ba3-55db-460a-9daa-d6e6e8b06766\") " Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.974177 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "06da071b-e1a1-42da-9e6e-fd957923876d" (UID: "06da071b-e1a1-42da-9e6e-fd957923876d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.979077 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "06da071b-e1a1-42da-9e6e-fd957923876d" (UID: "06da071b-e1a1-42da-9e6e-fd957923876d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.985546 4770 generic.go:334] "Generic (PLEG): container finished" podID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerID="ba22ef47e0582d9695bafcc9660a1b4adb815a65c5ef9b6fd33f435fbfb67a12" exitCode=0 Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.985616 4770 generic.go:334] "Generic (PLEG): container finished" podID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerID="3f304c4c9d4084df4db4c9d1ca2bc9d44285df6bb49699c82b5e8823e8e1ae3f" exitCode=143 Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.985691 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" event={"ID":"e328a4dc-7a4d-4cb0-82a8-f60df78328d8","Type":"ContainerDied","Data":"ba22ef47e0582d9695bafcc9660a1b4adb815a65c5ef9b6fd33f435fbfb67a12"} Oct 04 03:25:35 crc kubenswrapper[4770]: I1004 03:25:35.985724 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" event={"ID":"e328a4dc-7a4d-4cb0-82a8-f60df78328d8","Type":"ContainerDied","Data":"3f304c4c9d4084df4db4c9d1ca2bc9d44285df6bb49699c82b5e8823e8e1ae3f"} Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.025084 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-kube-api-access-rx8hd" (OuterVolumeSpecName: "kube-api-access-rx8hd") pod "06da071b-e1a1-42da-9e6e-fd957923876d" (UID: "06da071b-e1a1-42da-9e6e-fd957923876d"). InnerVolumeSpecName "kube-api-access-rx8hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.026751 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d236ba3-55db-460a-9daa-d6e6e8b06766-kube-api-access-phl9r" (OuterVolumeSpecName: "kube-api-access-phl9r") pod "1d236ba3-55db-460a-9daa-d6e6e8b06766" (UID: "1d236ba3-55db-460a-9daa-d6e6e8b06766"). InnerVolumeSpecName "kube-api-access-phl9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.028277 4770 generic.go:334] "Generic (PLEG): container finished" podID="d18c6351-6663-4a3c-84b6-2752f5e66e49" containerID="5bef3ded5c2c95e3979dba7e9d32dc622485969d7322f5de145044df6a55817c" exitCode=0 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.028397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementded7-account-delete-qxzt9" event={"ID":"d18c6351-6663-4a3c-84b6-2752f5e66e49","Type":"ContainerDied","Data":"5bef3ded5c2c95e3979dba7e9d32dc622485969d7322f5de145044df6a55817c"} Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.050427 4770 generic.go:334] "Generic (PLEG): container finished" podID="2af9166c-e2c5-462d-be52-340be8aadfd5" containerID="226e7dfe223b3700639cce5022b00883b758fc927a30ced5786418eeb4b2b69a" exitCode=0 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.050501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderec07-account-delete-j44sn" event={"ID":"2af9166c-e2c5-462d-be52-340be8aadfd5","Type":"ContainerDied","Data":"226e7dfe223b3700639cce5022b00883b758fc927a30ced5786418eeb4b2b69a"} Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.059811 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "06da071b-e1a1-42da-9e6e-fd957923876d" (UID: "06da071b-e1a1-42da-9e6e-fd957923876d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.061443 4770 generic.go:334] "Generic (PLEG): container finished" podID="e826d9bb-183e-401b-bc8b-6cfe0938a7f6" containerID="aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092" exitCode=0 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.061507 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e826d9bb-183e-401b-bc8b-6cfe0938a7f6","Type":"ContainerDied","Data":"aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092"} Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.086696 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhxhj\" (UniqueName: \"kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj\") pod \"novacell109d5-account-delete-6g4cl\" (UID: \"0f56d0a9-6e93-47b0-8ba2-4b04020f9296\") " pod="openstack/novacell109d5-account-delete-6g4cl" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.086912 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.086928 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx8hd\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-kube-api-access-rx8hd\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.086939 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06da071b-e1a1-42da-9e6e-fd957923876d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.086951 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phl9r\" (UniqueName: \"kubernetes.io/projected/1d236ba3-55db-460a-9daa-d6e6e8b06766-kube-api-access-phl9r\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.086963 4770 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06da071b-e1a1-42da-9e6e-fd957923876d-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.092487 4770 projected.go:194] Error preparing data for projected volume kube-api-access-qhxhj for pod openstack/novacell109d5-account-delete-6g4cl: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.092585 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj podName:0f56d0a9-6e93-47b0-8ba2-4b04020f9296 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:40.092558462 +0000 UTC m=+1351.384568174 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-qhxhj" (UniqueName: "kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj") pod "novacell109d5-account-delete-6g4cl" (UID: "0f56d0a9-6e93-47b0-8ba2-4b04020f9296") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.108437 4770 generic.go:334] "Generic (PLEG): container finished" podID="35f43e87-e3d9-4508-8bb1-4fe89c64017f" containerID="0218e06d6c52fad79e58c3f1c9b92b86ec1c9c65d56610da75164f0326b481a0" exitCode=0 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.108599 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"35f43e87-e3d9-4508-8bb1-4fe89c64017f","Type":"ContainerDied","Data":"0218e06d6c52fad79e58c3f1c9b92b86ec1c9c65d56610da75164f0326b481a0"} Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.195292 4770 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.195358 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:38.19533841 +0000 UTC m=+1349.487348122 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-config-data" not found Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.195428 4770 generic.go:334] "Generic (PLEG): container finished" podID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerID="ca61c9dff3e5c232108ba39a8b1a8f4f4f5b0eb23bc5838c1ee8c7e06106319e" exitCode=0 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.195467 4770 generic.go:334] "Generic (PLEG): container finished" podID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerID="ad1c6f5e9093316bec8bf19c97c384353c2014b68fb11557c892fae059e000b7" exitCode=143 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.195521 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66b9dbf78f-gv52v" event={"ID":"697bd5d5-22c6-4303-9077-434eaa099f4a","Type":"ContainerDied","Data":"ca61c9dff3e5c232108ba39a8b1a8f4f4f5b0eb23bc5838c1ee8c7e06106319e"} Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.195556 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66b9dbf78f-gv52v" event={"ID":"697bd5d5-22c6-4303-9077-434eaa099f4a","Type":"ContainerDied","Data":"ad1c6f5e9093316bec8bf19c97c384353c2014b68fb11557c892fae059e000b7"} Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.195706 4770 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.195784 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:38.195760982 +0000 UTC m=+1349.487770694 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-worker-config-data" not found Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.205754 4770 generic.go:334] "Generic (PLEG): container finished" podID="1d236ba3-55db-460a-9daa-d6e6e8b06766" containerID="f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8" exitCode=0 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.205915 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell109d5-account-delete-6g4cl" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.206897 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.207127 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1d236ba3-55db-460a-9daa-d6e6e8b06766","Type":"ContainerDied","Data":"f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8"} Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.207154 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1d236ba3-55db-460a-9daa-d6e6e8b06766","Type":"ContainerDied","Data":"4b1b87141144b8d9ae970d74e97c3a2a03a3f7e4342814bc56ce4a8bfb6fc71c"} Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.207283 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-576495ccf7-bslzp" podUID="8207573d-3535-4814-aff9-15c30c7600c7" containerName="barbican-worker-log" containerID="cri-o://6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc" gracePeriod=30 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.207430 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b8b845f-x9khq" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.207510 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-576495ccf7-bslzp" podUID="8207573d-3535-4814-aff9-15c30c7600c7" containerName="barbican-worker" containerID="cri-o://897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b" gracePeriod=30 Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.382687 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092 is running failed: container process not found" containerID="aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.383641 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092 is running failed: container process not found" containerID="aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.384059 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092 is running failed: container process not found" containerID="aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 03:25:36 crc kubenswrapper[4770]: E1004 03:25:36.384141 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="e826d9bb-183e-401b-bc8b-6cfe0938a7f6" containerName="nova-cell0-conductor-conductor" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.417846 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-config-data" (OuterVolumeSpecName: "config-data") pod "1d236ba3-55db-460a-9daa-d6e6e8b06766" (UID: "1d236ba3-55db-460a-9daa-d6e6e8b06766"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.518245 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.588628 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "1d236ba3-55db-460a-9daa-d6e6e8b06766" (UID: "1d236ba3-55db-460a-9daa-d6e6e8b06766"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.627521 4770 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.689060 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "1d236ba3-55db-460a-9daa-d6e6e8b06766" (UID: "1d236ba3-55db-460a-9daa-d6e6e8b06766"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.692149 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-config-data" (OuterVolumeSpecName: "config-data") pod "06da071b-e1a1-42da-9e6e-fd957923876d" (UID: "06da071b-e1a1-42da-9e6e-fd957923876d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.698550 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "06da071b-e1a1-42da-9e6e-fd957923876d" (UID: "06da071b-e1a1-42da-9e6e-fd957923876d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.724337 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d236ba3-55db-460a-9daa-d6e6e8b06766" (UID: "1d236ba3-55db-460a-9daa-d6e6e8b06766"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.724404 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-n8mcv"] Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.730924 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.730979 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.730998 4770 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d236ba3-55db-460a-9daa-d6e6e8b06766-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.739415 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.736378 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06da071b-e1a1-42da-9e6e-fd957923876d" (UID: "06da071b-e1a1-42da-9e6e-fd957923876d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.731969 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-n8mcv"] Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.753311 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi3321-account-delete-gmw9l"] Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.756511 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "06da071b-e1a1-42da-9e6e-fd957923876d" (UID: "06da071b-e1a1-42da-9e6e-fd957923876d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.771073 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3321-account-create-zxjrn"] Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.818225 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3321-account-create-zxjrn"] Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.841239 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.841655 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="ceilometer-central-agent" containerID="cri-o://3eef293d1922aeebed6a824d6768b260279f3c458835051702da305ac78ec1e4" gracePeriod=30 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.842241 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="proxy-httpd" containerID="cri-o://f6cc4fe8efa2daaa6113859b79bc85ffd8f259a17e76cdbe64107c051e5c6660" gracePeriod=30 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.842324 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="sg-core" containerID="cri-o://3ebcae43ad55173cedbd5766bb36e217f05f019084426c4bf34462bc4520a90f" gracePeriod=30 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.842364 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="ceilometer-notification-agent" containerID="cri-o://6fd96f4fbeb0043266b189a92e71e234f2a99b6b94cea0a2db77c70f4d686a5b" gracePeriod=30 Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.842529 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.842568 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da071b-e1a1-42da-9e6e-fd957923876d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.919084 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:25:36 crc kubenswrapper[4770]: I1004 03:25:36.919308 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="ae4b38bb-7bda-45e9-9d87-f79fe72e9008" containerName="kube-state-metrics" containerID="cri-o://136fdbb6c40c5913eff7c84a5bd828978d5112b321c3e6b34d37ea56010e8b07" gracePeriod=30 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.030723 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.031104 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="41309df4-53d7-42a8-8a5b-4816cc25fdfb" containerName="memcached" containerID="cri-o://5005796694c28178137720d1fa9150a9692d20f97c8d47bfb7d90339bebdf44b" gracePeriod=30 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.122544 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nq9d8"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.203111 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nq9d8"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.223235 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:36508->10.217.0.202:8775: read: connection reset by peer" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.223531 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:36522->10.217.0.202:8775: read: connection reset by peer" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.224646 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-29vw4"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.244063 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-29vw4"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252317 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystonef041-account-delete-d8qs9"] Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252773 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f70627-2bdd-4780-8044-75c3de4aee05" containerName="ovsdbserver-nb" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252791 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f70627-2bdd-4780-8044-75c3de4aee05" containerName="ovsdbserver-nb" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252814 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06da071b-e1a1-42da-9e6e-fd957923876d" containerName="proxy-server" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252821 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="06da071b-e1a1-42da-9e6e-fd957923876d" containerName="proxy-server" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252832 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerName="ovsdbserver-sb" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252838 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerName="ovsdbserver-sb" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252848 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c47d914-960a-4c5e-a6f3-584998910e09" containerName="init" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252854 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c47d914-960a-4c5e-a6f3-584998910e09" containerName="init" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252867 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerName="openstack-network-exporter" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252873 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerName="openstack-network-exporter" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252881 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f70627-2bdd-4780-8044-75c3de4aee05" containerName="openstack-network-exporter" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252887 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f70627-2bdd-4780-8044-75c3de4aee05" containerName="openstack-network-exporter" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252902 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4444038-6604-4e51-b27d-20d3616bf640" containerName="openstack-network-exporter" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252907 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4444038-6604-4e51-b27d-20d3616bf640" containerName="openstack-network-exporter" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252917 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06da071b-e1a1-42da-9e6e-fd957923876d" containerName="proxy-httpd" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252923 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="06da071b-e1a1-42da-9e6e-fd957923876d" containerName="proxy-httpd" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252938 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c47d914-960a-4c5e-a6f3-584998910e09" containerName="dnsmasq-dns" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252944 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c47d914-960a-4c5e-a6f3-584998910e09" containerName="dnsmasq-dns" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.252951 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d236ba3-55db-460a-9daa-d6e6e8b06766" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.252958 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d236ba3-55db-460a-9daa-d6e6e8b06766" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.253141 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="57f70627-2bdd-4780-8044-75c3de4aee05" containerName="openstack-network-exporter" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.253153 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4444038-6604-4e51-b27d-20d3616bf640" containerName="openstack-network-exporter" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.253161 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d236ba3-55db-460a-9daa-d6e6e8b06766" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.253171 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerName="ovsdbserver-sb" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.253190 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="06da071b-e1a1-42da-9e6e-fd957923876d" containerName="proxy-httpd" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.253201 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c47d914-960a-4c5e-a6f3-584998910e09" containerName="dnsmasq-dns" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.253213 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="57f70627-2bdd-4780-8044-75c3de4aee05" containerName="ovsdbserver-nb" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.253225 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="06da071b-e1a1-42da-9e6e-fd957923876d" containerName="proxy-server" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.253237 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4508feb-4ff8-47ec-86d7-7a9a0ad61a24" containerName="openstack-network-exporter" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.254081 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonef041-account-delete-d8qs9" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.262103 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.262174 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-866cc6bf5d-h867j"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.262344 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-866cc6bf5d-h867j" podUID="ee078c16-5858-4c24-b936-11d4aa568f65" containerName="keystone-api" containerID="cri-o://1f5f2c63939ced31d1a6828f3ba1395e0cbc60c93106d579d81b4050e2c43e88" gracePeriod=30 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.265181 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e826d9bb-183e-401b-bc8b-6cfe0938a7f6","Type":"ContainerDied","Data":"52a6d8bf39ba27e9c73e901e7a0cf29da2dad4207a79420d7a6b35d4610069d1"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.265204 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52a6d8bf39ba27e9c73e901e7a0cf29da2dad4207a79420d7a6b35d4610069d1" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.267078 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystonef041-account-delete-d8qs9"] Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.274253 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.289698 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.291241 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.291335 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="ovn-northd" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.293587 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-tqwql"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.298976 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" event={"ID":"9b1ac142-741d-48a1-bf5c-982bbf7bace6","Type":"ContainerStarted","Data":"a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.299198 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" podUID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerName="barbican-keystone-listener-log" containerID="cri-o://da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97" gracePeriod=30 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.299910 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" podUID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerName="barbican-keystone-listener" containerID="cri-o://a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1" gracePeriod=30 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.300000 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-tqwql"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.307467 4770 generic.go:334] "Generic (PLEG): container finished" podID="ae4b38bb-7bda-45e9-9d87-f79fe72e9008" containerID="136fdbb6c40c5913eff7c84a5bd828978d5112b321c3e6b34d37ea56010e8b07" exitCode=2 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.307548 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ae4b38bb-7bda-45e9-9d87-f79fe72e9008","Type":"ContainerDied","Data":"136fdbb6c40c5913eff7c84a5bd828978d5112b321c3e6b34d37ea56010e8b07"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.329228 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f041-account-create-dt88f"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.335145 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystonef041-account-delete-d8qs9"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.337795 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f041-account-create-dt88f"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.341222 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" podStartSLOduration=7.341210383 podStartE2EDuration="7.341210383s" podCreationTimestamp="2025-10-04 03:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:25:37.320345477 +0000 UTC m=+1348.612355189" watchObservedRunningTime="2025-10-04 03:25:37.341210383 +0000 UTC m=+1348.633220095" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.342709 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"35f43e87-e3d9-4508-8bb1-4fe89c64017f","Type":"ContainerDied","Data":"3f95e0d8695bd0efa45dc9005e6070c12f2bd8e488946031bf4b233e2927349d"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.342753 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f95e0d8695bd0efa45dc9005e6070c12f2bd8e488946031bf4b233e2927349d" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.344522 4770 generic.go:334] "Generic (PLEG): container finished" podID="87293d1f-5e33-4a9a-9978-67487fd14809" containerID="2ba074cd36ae8e2080bb5c06e842e5c1124a1cecb16c5a3f5f1ebd17d268f918" exitCode=0 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.344569 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87293d1f-5e33-4a9a-9978-67487fd14809","Type":"ContainerDied","Data":"2ba074cd36ae8e2080bb5c06e842e5c1124a1cecb16c5a3f5f1ebd17d268f918"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.344595 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87293d1f-5e33-4a9a-9978-67487fd14809","Type":"ContainerDied","Data":"a355b29233b80cd9d87bf672cfc5ce2d7d209b5e3ebe97a8765ab09f7cbb3c9e"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.344607 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a355b29233b80cd9d87bf672cfc5ce2d7d209b5e3ebe97a8765ab09f7cbb3c9e" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.361434 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66b9dbf78f-gv52v" event={"ID":"697bd5d5-22c6-4303-9077-434eaa099f4a","Type":"ContainerDied","Data":"4c6aed319f7675ffc676d7b125c8d46484cd776ab6d78a3724fd77ab7b9f489b"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.361486 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c6aed319f7675ffc676d7b125c8d46484cd776ab6d78a3724fd77ab7b9f489b" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.364442 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderec07-account-delete-j44sn" event={"ID":"2af9166c-e2c5-462d-be52-340be8aadfd5","Type":"ContainerDied","Data":"3409dbb112e2de95127ddfe61612a916b1c335555b22af005a35cb7cfe869fce"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.364467 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3409dbb112e2de95127ddfe61612a916b1c335555b22af005a35cb7cfe869fce" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.366921 4770 generic.go:334] "Generic (PLEG): container finished" podID="6d5328f9-e5e9-4ad5-96ff-02442117ff0e" containerID="2eb35e9ba346c7f87efe9e066071c98d4374aee1f72cd61b4a26a7bd7f997002" exitCode=0 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.366966 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3321-account-delete-gmw9l" event={"ID":"6d5328f9-e5e9-4ad5-96ff-02442117ff0e","Type":"ContainerDied","Data":"2eb35e9ba346c7f87efe9e066071c98d4374aee1f72cd61b4a26a7bd7f997002"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.368501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" event={"ID":"e328a4dc-7a4d-4cb0-82a8-f60df78328d8","Type":"ContainerDied","Data":"9cf55162da9e867dc8a8cde9e48d10d912ac61a3d2d08bbaa7e2df956c51e55b"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.368517 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cf55162da9e867dc8a8cde9e48d10d912ac61a3d2d08bbaa7e2df956c51e55b" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.369813 4770 generic.go:334] "Generic (PLEG): container finished" podID="76f9865e-be12-4821-a088-956061126f7b" containerID="2ed2b0c7307cdff4de6c68749e0b18dddacb71011729c291fbd02e14c6b6f756" exitCode=0 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.369846 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dbfd75b46-z42g2" event={"ID":"76f9865e-be12-4821-a088-956061126f7b","Type":"ContainerDied","Data":"2ed2b0c7307cdff4de6c68749e0b18dddacb71011729c291fbd02e14c6b6f756"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.370924 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementded7-account-delete-qxzt9" event={"ID":"d18c6351-6663-4a3c-84b6-2752f5e66e49","Type":"ContainerDied","Data":"5dddefeae5fd60c8df075cb5b7feee573fd182915531c9aad3bc15288fde970d"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.370939 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dddefeae5fd60c8df075cb5b7feee573fd182915531c9aad3bc15288fde970d" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.392582 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5hjg\" (UniqueName: \"kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg\") pod \"keystonef041-account-delete-d8qs9\" (UID: \"e13efcbd-d9b2-4c14-b681-471e7639d401\") " pod="openstack/keystonef041-account-delete-d8qs9" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.418311 4770 generic.go:334] "Generic (PLEG): container finished" podID="8207573d-3535-4814-aff9-15c30c7600c7" containerID="6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc" exitCode=143 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.418492 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-576495ccf7-bslzp" event={"ID":"8207573d-3535-4814-aff9-15c30c7600c7","Type":"ContainerDied","Data":"6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.428842 4770 generic.go:334] "Generic (PLEG): container finished" podID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerID="1f8470160c79b09e183b8b0818cd1ff1775ade9ca40792f2aa78671ca11f3a5b" exitCode=0 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.428926 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d","Type":"ContainerDied","Data":"1f8470160c79b09e183b8b0818cd1ff1775ade9ca40792f2aa78671ca11f3a5b"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.433421 4770 generic.go:334] "Generic (PLEG): container finished" podID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerID="4e30affad8c19685e28c648ecf8ff6c026d5dda9129c4944f788d2e762582b04" exitCode=0 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.433543 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d49edb80-7b9b-4b10-afbc-652ec540542e","Type":"ContainerDied","Data":"4e30affad8c19685e28c648ecf8ff6c026d5dda9129c4944f788d2e762582b04"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.435515 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f85c64c7b-jcqvq" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:53664->10.217.0.160:9311: read: connection reset by peer" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.435549 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f85c64c7b-jcqvq" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:53678->10.217.0.160:9311: read: connection reset by peer" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.448244 4770 generic.go:334] "Generic (PLEG): container finished" podID="c00a8476-9758-46f5-864f-9a7b80ca8bd1" containerID="9138162486e2a0c16b6c114810102c36c1c41126ad6f792e839348193d77a579" exitCode=0 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.448303 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronb0c9-account-delete-hc97s" event={"ID":"c00a8476-9758-46f5-864f-9a7b80ca8bd1","Type":"ContainerDied","Data":"9138162486e2a0c16b6c114810102c36c1c41126ad6f792e839348193d77a579"} Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.496970 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5hjg\" (UniqueName: \"kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg\") pod \"keystonef041-account-delete-d8qs9\" (UID: \"e13efcbd-d9b2-4c14-b681-471e7639d401\") " pod="openstack/keystonef041-account-delete-d8qs9" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.502479 4770 projected.go:194] Error preparing data for projected volume kube-api-access-h5hjg for pod openstack/keystonef041-account-delete-d8qs9: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.502585 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg podName:e13efcbd-d9b2-4c14-b681-471e7639d401 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:38.002554573 +0000 UTC m=+1349.294564285 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-h5hjg" (UniqueName: "kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg") pod "keystonef041-account-delete-d8qs9" (UID: "e13efcbd-d9b2-4c14-b681-471e7639d401") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.577562 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.624434 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.642039 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.642373 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.650935 4770 scope.go:117] "RemoveContainer" containerID="ed19534ae02538e57cfa8d6b65b5aaab360d08aa09ba030e130e63daca98294a" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.657404 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementded7-account-delete-qxzt9" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.664168 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell109d5-account-delete-6g4cl"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.670314 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell109d5-account-delete-6g4cl"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.673073 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" containerName="galera" containerID="cri-o://6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3" gracePeriod=30 Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.702306 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.709761 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data-custom\") pod \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710226 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-combined-ca-bundle\") pod \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710257 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data\") pod \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710284 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-generated\") pod \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710309 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-operator-scripts\") pod \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710338 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-combined-ca-bundle\") pod \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710595 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-logs\") pod \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710635 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kolla-config\") pod \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710668 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-secrets\") pod \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710782 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-default\") pod \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.710811 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw4g7\" (UniqueName: \"kubernetes.io/projected/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-kube-api-access-qw4g7\") pod \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\" (UID: \"e328a4dc-7a4d-4cb0-82a8-f60df78328d8\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.711069 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "35f43e87-e3d9-4508-8bb1-4fe89c64017f" (UID: "35f43e87-e3d9-4508-8bb1-4fe89c64017f"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.711850 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderec07-account-delete-j44sn" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.712731 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-logs" (OuterVolumeSpecName: "logs") pod "e328a4dc-7a4d-4cb0-82a8-f60df78328d8" (UID: "e328a4dc-7a4d-4cb0-82a8-f60df78328d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.713219 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gktjh\" (UniqueName: \"kubernetes.io/projected/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kube-api-access-gktjh\") pod \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.713322 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-galera-tls-certs\") pod \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\" (UID: \"35f43e87-e3d9-4508-8bb1-4fe89c64017f\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.713487 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "35f43e87-e3d9-4508-8bb1-4fe89c64017f" (UID: "35f43e87-e3d9-4508-8bb1-4fe89c64017f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.714448 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "35f43e87-e3d9-4508-8bb1-4fe89c64017f" (UID: "35f43e87-e3d9-4508-8bb1-4fe89c64017f"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.717371 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "35f43e87-e3d9-4508-8bb1-4fe89c64017f" (UID: "35f43e87-e3d9-4508-8bb1-4fe89c64017f"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.717782 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.717806 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.717816 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.717826 4770 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.717835 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/35f43e87-e3d9-4508-8bb1-4fe89c64017f-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.718860 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e328a4dc-7a4d-4cb0-82a8-f60df78328d8" (UID: "e328a4dc-7a4d-4cb0-82a8-f60df78328d8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.723782 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f56d0a9-6e93-47b0-8ba2-4b04020f9296" path="/var/lib/kubelet/pods/0f56d0a9-6e93-47b0-8ba2-4b04020f9296/volumes" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.724131 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19c6e72a-329c-4adc-8987-b227ff421e9e" path="/var/lib/kubelet/pods/19c6e72a-329c-4adc-8987-b227ff421e9e/volumes" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.724636 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="263fdaa5-3f50-4f4e-8908-6974cee72f0d" path="/var/lib/kubelet/pods/263fdaa5-3f50-4f4e-8908-6974cee72f0d/volumes" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.725122 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c9fa8cd-207f-4561-9116-ceea539098d8" path="/var/lib/kubelet/pods/3c9fa8cd-207f-4561-9116-ceea539098d8/volumes" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.726346 4770 scope.go:117] "RemoveContainer" containerID="5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.738962 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-kube-api-access-qw4g7" (OuterVolumeSpecName: "kube-api-access-qw4g7") pod "e328a4dc-7a4d-4cb0-82a8-f60df78328d8" (UID: "e328a4dc-7a4d-4cb0-82a8-f60df78328d8"). InnerVolumeSpecName "kube-api-access-qw4g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.739981 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-secrets" (OuterVolumeSpecName: "secrets") pod "35f43e87-e3d9-4508-8bb1-4fe89c64017f" (UID: "35f43e87-e3d9-4508-8bb1-4fe89c64017f"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.740124 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.744412 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-h5hjg], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystonef041-account-delete-d8qs9" podUID="e13efcbd-d9b2-4c14-b681-471e7639d401" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.745027 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45047928-b458-48d5-b62a-ab2bd91d6adb" path="/var/lib/kubelet/pods/45047928-b458-48d5-b62a-ab2bd91d6adb/volumes" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.745110 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.745696 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9dfa4b6-223e-447d-b62c-1f772c82802d" path="/var/lib/kubelet/pods/c9dfa4b6-223e-447d-b62c-1f772c82802d/volumes" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.746303 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4952962-0913-4c58-8647-9ff6823a2553" path="/var/lib/kubelet/pods/d4952962-0913-4c58-8647-9ff6823a2553/volumes" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.750131 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "mysql-db") pod "35f43e87-e3d9-4508-8bb1-4fe89c64017f" (UID: "35f43e87-e3d9-4508-8bb1-4fe89c64017f"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.756793 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.758850 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glanced8ca-account-delete-k2b6r" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.763521 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.763584 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" containerName="galera" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819270 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-scripts\") pod \"87293d1f-5e33-4a9a-9978-67487fd14809\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819366 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frwv7\" (UniqueName: \"kubernetes.io/projected/697bd5d5-22c6-4303-9077-434eaa099f4a-kube-api-access-frwv7\") pod \"697bd5d5-22c6-4303-9077-434eaa099f4a\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819426 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-combined-ca-bundle\") pod \"697bd5d5-22c6-4303-9077-434eaa099f4a\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819486 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-internal-tls-certs\") pod \"87293d1f-5e33-4a9a-9978-67487fd14809\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819564 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87293d1f-5e33-4a9a-9978-67487fd14809-logs\") pod \"87293d1f-5e33-4a9a-9978-67487fd14809\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819595 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data\") pod \"697bd5d5-22c6-4303-9077-434eaa099f4a\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819620 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data-custom\") pod \"87293d1f-5e33-4a9a-9978-67487fd14809\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819699 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-combined-ca-bundle\") pod \"87293d1f-5e33-4a9a-9978-67487fd14809\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819779 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh9vh\" (UniqueName: \"kubernetes.io/projected/87293d1f-5e33-4a9a-9978-67487fd14809-kube-api-access-gh9vh\") pod \"87293d1f-5e33-4a9a-9978-67487fd14809\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819817 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data\") pod \"87293d1f-5e33-4a9a-9978-67487fd14809\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819912 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/697bd5d5-22c6-4303-9077-434eaa099f4a-logs\") pod \"697bd5d5-22c6-4303-9077-434eaa099f4a\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.819964 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgnqx\" (UniqueName: \"kubernetes.io/projected/2af9166c-e2c5-462d-be52-340be8aadfd5-kube-api-access-wgnqx\") pod \"2af9166c-e2c5-462d-be52-340be8aadfd5\" (UID: \"2af9166c-e2c5-462d-be52-340be8aadfd5\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.820043 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87293d1f-5e33-4a9a-9978-67487fd14809-etc-machine-id\") pod \"87293d1f-5e33-4a9a-9978-67487fd14809\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.820094 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-config-data\") pod \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.820137 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-combined-ca-bundle\") pod \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.820182 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-public-tls-certs\") pod \"87293d1f-5e33-4a9a-9978-67487fd14809\" (UID: \"87293d1f-5e33-4a9a-9978-67487fd14809\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.820256 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzx4s\" (UniqueName: \"kubernetes.io/projected/d18c6351-6663-4a3c-84b6-2752f5e66e49-kube-api-access-kzx4s\") pod \"d18c6351-6663-4a3c-84b6-2752f5e66e49\" (UID: \"d18c6351-6663-4a3c-84b6-2752f5e66e49\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.820292 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data-custom\") pod \"697bd5d5-22c6-4303-9077-434eaa099f4a\" (UID: \"697bd5d5-22c6-4303-9077-434eaa099f4a\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.820362 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw9wr\" (UniqueName: \"kubernetes.io/projected/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-kube-api-access-xw9wr\") pod \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\" (UID: \"e826d9bb-183e-401b-bc8b-6cfe0938a7f6\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.821134 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhxhj\" (UniqueName: \"kubernetes.io/projected/0f56d0a9-6e93-47b0-8ba2-4b04020f9296-kube-api-access-qhxhj\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.821179 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.821212 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.821232 4770 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.821250 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw4g7\" (UniqueName: \"kubernetes.io/projected/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-kube-api-access-qw4g7\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.821258 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35f43e87-e3d9-4508-8bb1-4fe89c64017f" (UID: "35f43e87-e3d9-4508-8bb1-4fe89c64017f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.823126 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87293d1f-5e33-4a9a-9978-67487fd14809-logs" (OuterVolumeSpecName: "logs") pod "87293d1f-5e33-4a9a-9978-67487fd14809" (UID: "87293d1f-5e33-4a9a-9978-67487fd14809"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.832709 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/697bd5d5-22c6-4303-9077-434eaa099f4a-kube-api-access-frwv7" (OuterVolumeSpecName: "kube-api-access-frwv7") pod "697bd5d5-22c6-4303-9077-434eaa099f4a" (UID: "697bd5d5-22c6-4303-9077-434eaa099f4a"). InnerVolumeSpecName "kube-api-access-frwv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.835053 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-787bf4d777-x8gln"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.835090 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-787bf4d777-x8gln"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.842364 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/697bd5d5-22c6-4303-9077-434eaa099f4a-logs" (OuterVolumeSpecName: "logs") pod "697bd5d5-22c6-4303-9077-434eaa099f4a" (UID: "697bd5d5-22c6-4303-9077-434eaa099f4a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.842454 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87293d1f-5e33-4a9a-9978-67487fd14809-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "87293d1f-5e33-4a9a-9978-67487fd14809" (UID: "87293d1f-5e33-4a9a-9978-67487fd14809"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.871854 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d18c6351-6663-4a3c-84b6-2752f5e66e49-kube-api-access-kzx4s" (OuterVolumeSpecName: "kube-api-access-kzx4s") pod "d18c6351-6663-4a3c-84b6-2752f5e66e49" (UID: "d18c6351-6663-4a3c-84b6-2752f5e66e49"). InnerVolumeSpecName "kube-api-access-kzx4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.872335 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "87293d1f-5e33-4a9a-9978-67487fd14809" (UID: "87293d1f-5e33-4a9a-9978-67487fd14809"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.872542 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kube-api-access-gktjh" (OuterVolumeSpecName: "kube-api-access-gktjh") pod "35f43e87-e3d9-4508-8bb1-4fe89c64017f" (UID: "35f43e87-e3d9-4508-8bb1-4fe89c64017f"). InnerVolumeSpecName "kube-api-access-gktjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.878199 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87293d1f-5e33-4a9a-9978-67487fd14809-kube-api-access-gh9vh" (OuterVolumeSpecName: "kube-api-access-gh9vh") pod "87293d1f-5e33-4a9a-9978-67487fd14809" (UID: "87293d1f-5e33-4a9a-9978-67487fd14809"). InnerVolumeSpecName "kube-api-access-gh9vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.885303 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e328a4dc-7a4d-4cb0-82a8-f60df78328d8" (UID: "e328a4dc-7a4d-4cb0-82a8-f60df78328d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.885376 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.885456 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-scripts" (OuterVolumeSpecName: "scripts") pod "87293d1f-5e33-4a9a-9978-67487fd14809" (UID: "87293d1f-5e33-4a9a-9978-67487fd14809"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.904460 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-kube-api-access-xw9wr" (OuterVolumeSpecName: "kube-api-access-xw9wr") pod "e826d9bb-183e-401b-bc8b-6cfe0938a7f6" (UID: "e826d9bb-183e-401b-bc8b-6cfe0938a7f6"). InnerVolumeSpecName "kube-api-access-xw9wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.905081 4770 scope.go:117] "RemoveContainer" containerID="8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.905482 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "697bd5d5-22c6-4303-9077-434eaa099f4a" (UID: "697bd5d5-22c6-4303-9077-434eaa099f4a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.906051 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2af9166c-e2c5-462d-be52-340be8aadfd5-kube-api-access-wgnqx" (OuterVolumeSpecName: "kube-api-access-wgnqx") pod "2af9166c-e2c5-462d-be52-340be8aadfd5" (UID: "2af9166c-e2c5-462d-be52-340be8aadfd5"). InnerVolumeSpecName "kube-api-access-wgnqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.911295 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.923225 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjgk4\" (UniqueName: \"kubernetes.io/projected/d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c-kube-api-access-qjgk4\") pod \"d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c\" (UID: \"d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c\") " Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924693 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924713 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924724 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzx4s\" (UniqueName: \"kubernetes.io/projected/d18c6351-6663-4a3c-84b6-2752f5e66e49-kube-api-access-kzx4s\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924736 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924745 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw9wr\" (UniqueName: \"kubernetes.io/projected/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-kube-api-access-xw9wr\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924752 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924761 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frwv7\" (UniqueName: \"kubernetes.io/projected/697bd5d5-22c6-4303-9077-434eaa099f4a-kube-api-access-frwv7\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924771 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87293d1f-5e33-4a9a-9978-67487fd14809-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924779 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924787 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gktjh\" (UniqueName: \"kubernetes.io/projected/35f43e87-e3d9-4508-8bb1-4fe89c64017f-kube-api-access-gktjh\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924796 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh9vh\" (UniqueName: \"kubernetes.io/projected/87293d1f-5e33-4a9a-9978-67487fd14809-kube-api-access-gh9vh\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924804 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/697bd5d5-22c6-4303-9077-434eaa099f4a-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924812 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgnqx\" (UniqueName: \"kubernetes.io/projected/2af9166c-e2c5-462d-be52-340be8aadfd5-kube-api-access-wgnqx\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.924822 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87293d1f-5e33-4a9a-9978-67487fd14809-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.949672 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c-kube-api-access-qjgk4" (OuterVolumeSpecName: "kube-api-access-qjgk4") pod "d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c" (UID: "d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c"). InnerVolumeSpecName "kube-api-access-qjgk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.951272 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-x9khq"] Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.952141 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e826d9bb-183e-401b-bc8b-6cfe0938a7f6" (UID: "e826d9bb-183e-401b-bc8b-6cfe0938a7f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.978982 4770 scope.go:117] "RemoveContainer" containerID="5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7" Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.979530 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7\": container with ID starting with 5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7 not found: ID does not exist" containerID="5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.979589 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7"} err="failed to get container status \"5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7\": rpc error: code = NotFound desc = could not find container \"5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7\": container with ID starting with 5d2ae5772480064eed69a2a86bd83ec2d5263d325514fd5f6c65f9f45dea27a7 not found: ID does not exist" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.979619 4770 scope.go:117] "RemoveContainer" containerID="8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.979780 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c5b8b845f-x9khq"] Oct 04 03:25:37 crc kubenswrapper[4770]: E1004 03:25:37.979886 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70\": container with ID starting with 8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70 not found: ID does not exist" containerID="8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.980307 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70"} err="failed to get container status \"8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70\": rpc error: code = NotFound desc = could not find container \"8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70\": container with ID starting with 8187b0968098201be2151b069de465985e0d44fe0a6a6e74b8367ff464709d70 not found: ID does not exist" Oct 04 03:25:37 crc kubenswrapper[4770]: I1004 03:25:37.980444 4770 scope.go:117] "RemoveContainer" containerID="f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.033186 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5hjg\" (UniqueName: \"kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg\") pod \"keystonef041-account-delete-d8qs9\" (UID: \"e13efcbd-d9b2-4c14-b681-471e7639d401\") " pod="openstack/keystonef041-account-delete-d8qs9" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.033370 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjgk4\" (UniqueName: \"kubernetes.io/projected/d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c-kube-api-access-qjgk4\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.033387 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.035145 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87293d1f-5e33-4a9a-9978-67487fd14809" (UID: "87293d1f-5e33-4a9a-9978-67487fd14809"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.037995 4770 projected.go:194] Error preparing data for projected volume kube-api-access-h5hjg for pod openstack/keystonef041-account-delete-d8qs9: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.039580 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg podName:e13efcbd-d9b2-4c14-b681-471e7639d401 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:39.039556469 +0000 UTC m=+1350.331566191 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-h5hjg" (UniqueName: "kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg") pod "keystonef041-account-delete-d8qs9" (UID: "e13efcbd-d9b2-4c14-b681-471e7639d401") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.049786 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-config-data" (OuterVolumeSpecName: "config-data") pod "e826d9bb-183e-401b-bc8b-6cfe0938a7f6" (UID: "e826d9bb-183e-401b-bc8b-6cfe0938a7f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.087342 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "35f43e87-e3d9-4508-8bb1-4fe89c64017f" (UID: "35f43e87-e3d9-4508-8bb1-4fe89c64017f"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.105020 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.127951 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data" (OuterVolumeSpecName: "config-data") pod "697bd5d5-22c6-4303-9077-434eaa099f4a" (UID: "697bd5d5-22c6-4303-9077-434eaa099f4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.149628 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.149675 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.149691 4770 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/35f43e87-e3d9-4508-8bb1-4fe89c64017f-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.149715 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.149728 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e826d9bb-183e-401b-bc8b-6cfe0938a7f6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.222382 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="41309df4-53d7-42a8-8a5b-4816cc25fdfb" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.106:11211: connect: connection refused" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.236869 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "697bd5d5-22c6-4303-9077-434eaa099f4a" (UID: "697bd5d5-22c6-4303-9077-434eaa099f4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.236993 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "87293d1f-5e33-4a9a-9978-67487fd14809" (UID: "87293d1f-5e33-4a9a-9978-67487fd14809"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.237565 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data" (OuterVolumeSpecName: "config-data") pod "87293d1f-5e33-4a9a-9978-67487fd14809" (UID: "87293d1f-5e33-4a9a-9978-67487fd14809"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.253683 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.253730 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.253745 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697bd5d5-22c6-4303-9077-434eaa099f4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.253749 4770 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.253841 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:42.253816993 +0000 UTC m=+1353.545826795 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-worker-config-data" not found Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.253955 4770 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.254121 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:42.254090971 +0000 UTC m=+1353.546100683 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-config-data" not found Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.257041 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data" (OuterVolumeSpecName: "config-data") pod "e328a4dc-7a4d-4cb0-82a8-f60df78328d8" (UID: "e328a4dc-7a4d-4cb0-82a8-f60df78328d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.274875 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "87293d1f-5e33-4a9a-9978-67487fd14809" (UID: "87293d1f-5e33-4a9a-9978-67487fd14809"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.362318 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87293d1f-5e33-4a9a-9978-67487fd14809-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.362358 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e328a4dc-7a4d-4cb0-82a8-f60df78328d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.469491 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronb0c9-account-delete-hc97s" event={"ID":"c00a8476-9758-46f5-864f-9a7b80ca8bd1","Type":"ContainerDied","Data":"0059ba19a9d39f65a573385056c75a8ed13f8e1072497933e7398653b1dac84e"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.469532 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0059ba19a9d39f65a573385056c75a8ed13f8e1072497933e7398653b1dac84e" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.471308 4770 generic.go:334] "Generic (PLEG): container finished" podID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerID="f6cc4fe8efa2daaa6113859b79bc85ffd8f259a17e76cdbe64107c051e5c6660" exitCode=0 Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.471329 4770 generic.go:334] "Generic (PLEG): container finished" podID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerID="3ebcae43ad55173cedbd5766bb36e217f05f019084426c4bf34462bc4520a90f" exitCode=2 Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.471337 4770 generic.go:334] "Generic (PLEG): container finished" podID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerID="3eef293d1922aeebed6a824d6768b260279f3c458835051702da305ac78ec1e4" exitCode=0 Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.471398 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerDied","Data":"f6cc4fe8efa2daaa6113859b79bc85ffd8f259a17e76cdbe64107c051e5c6660"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.471415 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerDied","Data":"3ebcae43ad55173cedbd5766bb36e217f05f019084426c4bf34462bc4520a90f"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.471424 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerDied","Data":"3eef293d1922aeebed6a824d6768b260279f3c458835051702da305ac78ec1e4"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.473246 4770 generic.go:334] "Generic (PLEG): container finished" podID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerID="da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97" exitCode=143 Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.473290 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" event={"ID":"9b1ac142-741d-48a1-bf5c-982bbf7bace6","Type":"ContainerDied","Data":"da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.474951 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d","Type":"ContainerDied","Data":"ae4f904c53b0cb037ba79bc489ee1fcf211d72c4af3c433201d4c6599fddff11"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.474974 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae4f904c53b0cb037ba79bc489ee1fcf211d72c4af3c433201d4c6599fddff11" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.497734 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d49edb80-7b9b-4b10-afbc-652ec540542e","Type":"ContainerDied","Data":"d6066afec7af3ac77155b530eef0a09aaada0e0f6fb17bf3d05343aaad18f073"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.497778 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6066afec7af3ac77155b530eef0a09aaada0e0f6fb17bf3d05343aaad18f073" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.499113 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dbfd75b46-z42g2" event={"ID":"76f9865e-be12-4821-a088-956061126f7b","Type":"ContainerDied","Data":"b6021757cffcd01c59af1172ba928c987a4ca81d21ca51c80b67cb2beed40a6e"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.499134 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6021757cffcd01c59af1172ba928c987a4ca81d21ca51c80b67cb2beed40a6e" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.500161 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanced8ca-account-delete-k2b6r" event={"ID":"d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c","Type":"ContainerDied","Data":"10a76a22d6db2361ded091227c2fec250e401a12fcc957558c9406bf29541919"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.500280 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glanced8ca-account-delete-k2b6r" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.511610 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ae4b38bb-7bda-45e9-9d87-f79fe72e9008","Type":"ContainerDied","Data":"0d0dfd421a79a3b41e1ee5746d2e20b78b0aec53f57e39060355d20191138138"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.511675 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d0dfd421a79a3b41e1ee5746d2e20b78b0aec53f57e39060355d20191138138" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.514918 4770 generic.go:334] "Generic (PLEG): container finished" podID="41309df4-53d7-42a8-8a5b-4816cc25fdfb" containerID="5005796694c28178137720d1fa9150a9692d20f97c8d47bfb7d90339bebdf44b" exitCode=0 Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.515029 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"41309df4-53d7-42a8-8a5b-4816cc25fdfb","Type":"ContainerDied","Data":"5005796694c28178137720d1fa9150a9692d20f97c8d47bfb7d90339bebdf44b"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.529725 4770 generic.go:334] "Generic (PLEG): container finished" podID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerID="6bba8b83577d71e684e12c6e57b942c931915975b422c9aa9892702880a260fc" exitCode=0 Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.529826 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b795408-0144-4d73-8a56-043ce3e38ee6","Type":"ContainerDied","Data":"6bba8b83577d71e684e12c6e57b942c931915975b422c9aa9892702880a260fc"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.529865 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b795408-0144-4d73-8a56-043ce3e38ee6","Type":"ContainerDied","Data":"c686cdba7b99d359ef3efed5230a52bf695646bcd700ca78bdc38e78274cf6f7"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.529880 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c686cdba7b99d359ef3efed5230a52bf695646bcd700ca78bdc38e78274cf6f7" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.543720 4770 generic.go:334] "Generic (PLEG): container finished" podID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerID="f652f94cc4c1e2df6ea6011817891a4c91174bbbbee764ff5579bda21b7ab8fb" exitCode=0 Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.543772 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f85c64c7b-jcqvq" event={"ID":"83d745cd-57f5-4624-9d2c-75353e52fa56","Type":"ContainerDied","Data":"f652f94cc4c1e2df6ea6011817891a4c91174bbbbee764ff5579bda21b7ab8fb"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.543838 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f85c64c7b-jcqvq" event={"ID":"83d745cd-57f5-4624-9d2c-75353e52fa56","Type":"ContainerDied","Data":"3fc97d22a0ca2a02e0e8c90826afa1c5686946f611beac9c19eacf4fb2e9a19d"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.543856 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fc97d22a0ca2a02e0e8c90826afa1c5686946f611beac9c19eacf4fb2e9a19d" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.546123 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.552490 4770 generic.go:334] "Generic (PLEG): container finished" podID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerID="fa7c4e0199de7a74a6a51a933f598585e135eac2ee7b99b1d3d49fa44c55ad5a" exitCode=0 Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.552653 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66b9dbf78f-gv52v" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.552956 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonef041-account-delete-d8qs9" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.553057 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79ab7038-39be-4c09-82ca-00f83aa69e37","Type":"ContainerDied","Data":"fa7c4e0199de7a74a6a51a933f598585e135eac2ee7b99b1d3d49fa44c55ad5a"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.553094 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"79ab7038-39be-4c09-82ca-00f83aa69e37","Type":"ContainerDied","Data":"bdec7769a483ec499ca82ac6def167d67b14c965f9ed69c0964ade9fd5a29f1b"} Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.553112 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdec7769a483ec499ca82ac6def167d67b14c965f9ed69c0964ade9fd5a29f1b" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.553251 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderec07-account-delete-j44sn" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.555122 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.555357 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.556862 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5877f57886-5xk9s" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.557290 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementded7-account-delete-qxzt9" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.559150 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.567323 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glanced8ca-account-delete-k2b6r"] Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.574473 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.575151 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.576946 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.576980 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.577961 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.578061 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glanced8ca-account-delete-k2b6r"] Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.579938 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.589203 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.595871 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.600855 4770 scope.go:117] "RemoveContainer" containerID="f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8" Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.601968 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8\": container with ID starting with f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8 not found: ID does not exist" containerID="f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.602029 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8"} err="failed to get container status \"f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8\": rpc error: code = NotFound desc = could not find container \"f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8\": container with ID starting with f4512dbce1f841290649fcc4a0876883d090b0b931da6d0686df833b512ad2e8 not found: ID does not exist" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.602065 4770 scope.go:117] "RemoveContainer" containerID="4435c5f48a844ce2bd51729a6c5ff1242e1229d3621dec9c5c639c58455cb623" Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.605257 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.605334 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.617514 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667413 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-httpd-run\") pod \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667589 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-config-data\") pod \"76f9865e-be12-4821-a088-956061126f7b\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667632 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-public-tls-certs\") pod \"76f9865e-be12-4821-a088-956061126f7b\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667690 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-combined-ca-bundle\") pod \"76f9865e-be12-4821-a088-956061126f7b\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667757 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-config-data\") pod \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667815 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-logs\") pod \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667895 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-scripts\") pod \"76f9865e-be12-4821-a088-956061126f7b\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667917 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-scripts\") pod \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667936 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-internal-tls-certs\") pod \"76f9865e-be12-4821-a088-956061126f7b\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667955 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-public-tls-certs\") pod \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.667983 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f9865e-be12-4821-a088-956061126f7b-logs\") pod \"76f9865e-be12-4821-a088-956061126f7b\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.668020 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-797p7\" (UniqueName: \"kubernetes.io/projected/76f9865e-be12-4821-a088-956061126f7b-kube-api-access-797p7\") pod \"76f9865e-be12-4821-a088-956061126f7b\" (UID: \"76f9865e-be12-4821-a088-956061126f7b\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.668042 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-combined-ca-bundle\") pod \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.668129 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.668179 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9sdf\" (UniqueName: \"kubernetes.io/projected/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-kube-api-access-w9sdf\") pod \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\" (UID: \"1cb81588-b9f4-4d8a-8baf-33aebaa6e64d\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.668863 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-logs" (OuterVolumeSpecName: "logs") pod "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" (UID: "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.669283 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" (UID: "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.672412 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76f9865e-be12-4821-a088-956061126f7b-logs" (OuterVolumeSpecName: "logs") pod "76f9865e-be12-4821-a088-956061126f7b" (UID: "76f9865e-be12-4821-a088-956061126f7b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.677706 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76f9865e-be12-4821-a088-956061126f7b-kube-api-access-797p7" (OuterVolumeSpecName: "kube-api-access-797p7") pod "76f9865e-be12-4821-a088-956061126f7b" (UID: "76f9865e-be12-4821-a088-956061126f7b"). InnerVolumeSpecName "kube-api-access-797p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.678292 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-kube-api-access-w9sdf" (OuterVolumeSpecName: "kube-api-access-w9sdf") pod "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" (UID: "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d"). InnerVolumeSpecName "kube-api-access-w9sdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.685791 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.701150 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-scripts" (OuterVolumeSpecName: "scripts") pod "76f9865e-be12-4821-a088-956061126f7b" (UID: "76f9865e-be12-4821-a088-956061126f7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.702854 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" (UID: "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.703864 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-scripts" (OuterVolumeSpecName: "scripts") pod "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" (UID: "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.709347 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb0c9-account-delete-hc97s" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.727160 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" (UID: "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.749456 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-config-data" (OuterVolumeSpecName: "config-data") pod "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" (UID: "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.751488 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.757049 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-c2v9x" podUID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" containerName="ovn-controller" probeResult="failure" output="command timed out" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769341 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76f9865e-be12-4821-a088-956061126f7b" (UID: "76f9865e-be12-4821-a088-956061126f7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769508 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-config\") pod \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769583 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-combined-ca-bundle\") pod \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769624 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-scripts\") pod \"d49edb80-7b9b-4b10-afbc-652ec540542e\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769652 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-config-data\") pod \"d49edb80-7b9b-4b10-afbc-652ec540542e\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769699 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-internal-tls-certs\") pod \"d49edb80-7b9b-4b10-afbc-652ec540542e\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769735 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-httpd-run\") pod \"d49edb80-7b9b-4b10-afbc-652ec540542e\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769803 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lmb9\" (UniqueName: \"kubernetes.io/projected/d49edb80-7b9b-4b10-afbc-652ec540542e-kube-api-access-5lmb9\") pod \"d49edb80-7b9b-4b10-afbc-652ec540542e\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769870 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-combined-ca-bundle\") pod \"d49edb80-7b9b-4b10-afbc-652ec540542e\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769922 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99r7b\" (UniqueName: \"kubernetes.io/projected/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-api-access-99r7b\") pod \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769956 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-certs\") pod \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\" (UID: \"ae4b38bb-7bda-45e9-9d87-f79fe72e9008\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769974 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-logs\") pod \"d49edb80-7b9b-4b10-afbc-652ec540542e\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.769999 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"d49edb80-7b9b-4b10-afbc-652ec540542e\" (UID: \"d49edb80-7b9b-4b10-afbc-652ec540542e\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775277 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775509 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9sdf\" (UniqueName: \"kubernetes.io/projected/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-kube-api-access-w9sdf\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775523 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775533 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775541 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775552 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775567 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775576 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775584 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f9865e-be12-4821-a088-956061126f7b-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775594 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-797p7\" (UniqueName: \"kubernetes.io/projected/76f9865e-be12-4821-a088-956061126f7b-kube-api-access-797p7\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.775603 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.777067 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-logs" (OuterVolumeSpecName: "logs") pod "d49edb80-7b9b-4b10-afbc-652ec540542e" (UID: "d49edb80-7b9b-4b10-afbc-652ec540542e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.777344 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "d49edb80-7b9b-4b10-afbc-652ec540542e" (UID: "d49edb80-7b9b-4b10-afbc-652ec540542e"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.777606 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.777655 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data podName:2def82a8-5927-4ebe-ac87-e8ad232797ee nodeName:}" failed. No retries permitted until 2025-10-04 03:25:46.777638996 +0000 UTC m=+1358.069648708 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data") pod "rabbitmq-server-0" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee") : configmap "rabbitmq-config-data" not found Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.778035 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.785776 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-api-access-99r7b" (OuterVolumeSpecName: "kube-api-access-99r7b") pod "ae4b38bb-7bda-45e9-9d87-f79fe72e9008" (UID: "ae4b38bb-7bda-45e9-9d87-f79fe72e9008"). InnerVolumeSpecName "kube-api-access-99r7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.789474 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderec07-account-delete-j44sn"] Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.795461 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d49edb80-7b9b-4b10-afbc-652ec540542e" (UID: "d49edb80-7b9b-4b10-afbc-652ec540542e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.810733 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.815922 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonef041-account-delete-d8qs9" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.815925 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-scripts" (OuterVolumeSpecName: "scripts") pod "d49edb80-7b9b-4b10-afbc-652ec540542e" (UID: "d49edb80-7b9b-4b10-afbc-652ec540542e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.835466 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49edb80-7b9b-4b10-afbc-652ec540542e-kube-api-access-5lmb9" (OuterVolumeSpecName: "kube-api-access-5lmb9") pod "d49edb80-7b9b-4b10-afbc-652ec540542e" (UID: "d49edb80-7b9b-4b10-afbc-652ec540542e"). InnerVolumeSpecName "kube-api-access-5lmb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.842250 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.842519 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinderec07-account-delete-j44sn"] Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.845799 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.853055 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 03:25:38 crc kubenswrapper[4770]: E1004 03:25:38.853209 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="ef58a910-f85e-4e14-95b5-807741097485" containerName="nova-cell1-conductor-conductor" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.857190 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.877897 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-public-tls-certs\") pod \"4b795408-0144-4d73-8a56-043ce3e38ee6\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.877961 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nppcd\" (UniqueName: \"kubernetes.io/projected/79ab7038-39be-4c09-82ca-00f83aa69e37-kube-api-access-nppcd\") pod \"79ab7038-39be-4c09-82ca-00f83aa69e37\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878049 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-nova-metadata-tls-certs\") pod \"79ab7038-39be-4c09-82ca-00f83aa69e37\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878083 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d745cd-57f5-4624-9d2c-75353e52fa56-logs\") pod \"83d745cd-57f5-4624-9d2c-75353e52fa56\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878166 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8v9f\" (UniqueName: \"kubernetes.io/projected/83d745cd-57f5-4624-9d2c-75353e52fa56-kube-api-access-t8v9f\") pod \"83d745cd-57f5-4624-9d2c-75353e52fa56\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878185 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-config-data\") pod \"4b795408-0144-4d73-8a56-043ce3e38ee6\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878239 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data\") pod \"83d745cd-57f5-4624-9d2c-75353e52fa56\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878299 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-config-data\") pod \"79ab7038-39be-4c09-82ca-00f83aa69e37\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878316 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-internal-tls-certs\") pod \"4b795408-0144-4d73-8a56-043ce3e38ee6\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878335 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nblt4\" (UniqueName: \"kubernetes.io/projected/c00a8476-9758-46f5-864f-9a7b80ca8bd1-kube-api-access-nblt4\") pod \"c00a8476-9758-46f5-864f-9a7b80ca8bd1\" (UID: \"c00a8476-9758-46f5-864f-9a7b80ca8bd1\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878368 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-public-tls-certs\") pod \"83d745cd-57f5-4624-9d2c-75353e52fa56\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878395 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ab7038-39be-4c09-82ca-00f83aa69e37-logs\") pod \"79ab7038-39be-4c09-82ca-00f83aa69e37\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878481 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-combined-ca-bundle\") pod \"4b795408-0144-4d73-8a56-043ce3e38ee6\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878504 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-internal-tls-certs\") pod \"83d745cd-57f5-4624-9d2c-75353e52fa56\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878527 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data-custom\") pod \"83d745cd-57f5-4624-9d2c-75353e52fa56\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878546 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b795408-0144-4d73-8a56-043ce3e38ee6-logs\") pod \"4b795408-0144-4d73-8a56-043ce3e38ee6\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878569 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98pqk\" (UniqueName: \"kubernetes.io/projected/4b795408-0144-4d73-8a56-043ce3e38ee6-kube-api-access-98pqk\") pod \"4b795408-0144-4d73-8a56-043ce3e38ee6\" (UID: \"4b795408-0144-4d73-8a56-043ce3e38ee6\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878587 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-combined-ca-bundle\") pod \"79ab7038-39be-4c09-82ca-00f83aa69e37\" (UID: \"79ab7038-39be-4c09-82ca-00f83aa69e37\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878611 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-combined-ca-bundle\") pod \"83d745cd-57f5-4624-9d2c-75353e52fa56\" (UID: \"83d745cd-57f5-4624-9d2c-75353e52fa56\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.878989 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99r7b\" (UniqueName: \"kubernetes.io/projected/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-api-access-99r7b\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.879115 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.879141 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.879150 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.879159 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.879168 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d49edb80-7b9b-4b10-afbc-652ec540542e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.879178 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lmb9\" (UniqueName: \"kubernetes.io/projected/d49edb80-7b9b-4b10-afbc-652ec540542e-kube-api-access-5lmb9\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.890459 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae4b38bb-7bda-45e9-9d87-f79fe72e9008" (UID: "ae4b38bb-7bda-45e9-9d87-f79fe72e9008"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.896855 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b795408-0144-4d73-8a56-043ce3e38ee6-logs" (OuterVolumeSpecName: "logs") pod "4b795408-0144-4d73-8a56-043ce3e38ee6" (UID: "4b795408-0144-4d73-8a56-043ce3e38ee6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.897073 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79ab7038-39be-4c09-82ca-00f83aa69e37-logs" (OuterVolumeSpecName: "logs") pod "79ab7038-39be-4c09-82ca-00f83aa69e37" (UID: "79ab7038-39be-4c09-82ca-00f83aa69e37"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.900598 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.902200 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83d745cd-57f5-4624-9d2c-75353e52fa56-logs" (OuterVolumeSpecName: "logs") pod "83d745cd-57f5-4624-9d2c-75353e52fa56" (UID: "83d745cd-57f5-4624-9d2c-75353e52fa56"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.902543 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c00a8476-9758-46f5-864f-9a7b80ca8bd1-kube-api-access-nblt4" (OuterVolumeSpecName: "kube-api-access-nblt4") pod "c00a8476-9758-46f5-864f-9a7b80ca8bd1" (UID: "c00a8476-9758-46f5-864f-9a7b80ca8bd1"). InnerVolumeSpecName "kube-api-access-nblt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.904628 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-c2v9x" podUID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" containerName="ovn-controller" probeResult="failure" output=< Oct 04 03:25:38 crc kubenswrapper[4770]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Oct 04 03:25:38 crc kubenswrapper[4770]: > Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.909821 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79ab7038-39be-4c09-82ca-00f83aa69e37-kube-api-access-nppcd" (OuterVolumeSpecName: "kube-api-access-nppcd") pod "79ab7038-39be-4c09-82ca-00f83aa69e37" (UID: "79ab7038-39be-4c09-82ca-00f83aa69e37"). InnerVolumeSpecName "kube-api-access-nppcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.931906 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.948246 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.962231 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" (UID: "1cb81588-b9f4-4d8a-8baf-33aebaa6e64d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.962319 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b795408-0144-4d73-8a56-043ce3e38ee6-kube-api-access-98pqk" (OuterVolumeSpecName: "kube-api-access-98pqk") pod "4b795408-0144-4d73-8a56-043ce3e38ee6" (UID: "4b795408-0144-4d73-8a56-043ce3e38ee6"). InnerVolumeSpecName "kube-api-access-98pqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.962384 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83d745cd-57f5-4624-9d2c-75353e52fa56-kube-api-access-t8v9f" (OuterVolumeSpecName: "kube-api-access-t8v9f") pod "83d745cd-57f5-4624-9d2c-75353e52fa56" (UID: "83d745cd-57f5-4624-9d2c-75353e52fa56"). InnerVolumeSpecName "kube-api-access-t8v9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.965177 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "83d745cd-57f5-4624-9d2c-75353e52fa56" (UID: "83d745cd-57f5-4624-9d2c-75353e52fa56"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.966466 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.976965 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.982939 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq24q\" (UniqueName: \"kubernetes.io/projected/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kube-api-access-rq24q\") pod \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.983027 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-config-data\") pod \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.983083 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kolla-config\") pod \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.983268 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-memcached-tls-certs\") pod \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.983295 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-combined-ca-bundle\") pod \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\" (UID: \"41309df4-53d7-42a8-8a5b-4816cc25fdfb\") " Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.983829 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nppcd\" (UniqueName: \"kubernetes.io/projected/79ab7038-39be-4c09-82ca-00f83aa69e37-kube-api-access-nppcd\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.983851 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d745cd-57f5-4624-9d2c-75353e52fa56-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.984000 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8v9f\" (UniqueName: \"kubernetes.io/projected/83d745cd-57f5-4624-9d2c-75353e52fa56-kube-api-access-t8v9f\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.984024 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.984033 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.984042 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nblt4\" (UniqueName: \"kubernetes.io/projected/c00a8476-9758-46f5-864f-9a7b80ca8bd1-kube-api-access-nblt4\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.984055 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.984065 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79ab7038-39be-4c09-82ca-00f83aa69e37-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.984073 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.984082 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b795408-0144-4d73-8a56-043ce3e38ee6-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.984091 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98pqk\" (UniqueName: \"kubernetes.io/projected/4b795408-0144-4d73-8a56-043ce3e38ee6-kube-api-access-98pqk\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.986406 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-config-data" (OuterVolumeSpecName: "config-data") pod "41309df4-53d7-42a8-8a5b-4816cc25fdfb" (UID: "41309df4-53d7-42a8-8a5b-4816cc25fdfb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:38 crc kubenswrapper[4770]: I1004 03:25:38.986740 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "41309df4-53d7-42a8-8a5b-4816cc25fdfb" (UID: "41309df4-53d7-42a8-8a5b-4816cc25fdfb"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.014219 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5877f57886-5xk9s"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.045222 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-5877f57886-5xk9s"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.052735 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kube-api-access-rq24q" (OuterVolumeSpecName: "kube-api-access-rq24q") pod "41309df4-53d7-42a8-8a5b-4816cc25fdfb" (UID: "41309df4-53d7-42a8-8a5b-4816cc25fdfb"). InnerVolumeSpecName "kube-api-access-rq24q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.052831 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-66b9dbf78f-gv52v"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.084664 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d49edb80-7b9b-4b10-afbc-652ec540542e" (UID: "d49edb80-7b9b-4b10-afbc-652ec540542e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.084987 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d49edb80-7b9b-4b10-afbc-652ec540542e" (UID: "d49edb80-7b9b-4b10-afbc-652ec540542e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.086389 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5hjg\" (UniqueName: \"kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg\") pod \"keystonef041-account-delete-d8qs9\" (UID: \"e13efcbd-d9b2-4c14-b681-471e7639d401\") " pod="openstack/keystonef041-account-delete-d8qs9" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.086721 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.086811 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq24q\" (UniqueName: \"kubernetes.io/projected/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kube-api-access-rq24q\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.086888 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.086961 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.087105 4770 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41309df4-53d7-42a8-8a5b-4816cc25fdfb-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: E1004 03:25:39.093436 4770 projected.go:194] Error preparing data for projected volume kube-api-access-h5hjg for pod openstack/keystonef041-account-delete-d8qs9: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 04 03:25:39 crc kubenswrapper[4770]: E1004 03:25:39.093497 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg podName:e13efcbd-d9b2-4c14-b681-471e7639d401 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:41.093480156 +0000 UTC m=+1352.385489868 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-h5hjg" (UniqueName: "kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg") pod "keystonef041-account-delete-d8qs9" (UID: "e13efcbd-d9b2-4c14-b681-471e7639d401") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.093766 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-66b9dbf78f-gv52v"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.096885 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-config-data" (OuterVolumeSpecName: "config-data") pod "76f9865e-be12-4821-a088-956061126f7b" (UID: "76f9865e-be12-4821-a088-956061126f7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.101425 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementded7-account-delete-qxzt9"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.113281 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementded7-account-delete-qxzt9"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.122079 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "ae4b38bb-7bda-45e9-9d87-f79fe72e9008" (UID: "ae4b38bb-7bda-45e9-9d87-f79fe72e9008"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.166677 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79ab7038-39be-4c09-82ca-00f83aa69e37" (UID: "79ab7038-39be-4c09-82ca-00f83aa69e37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.189644 4770 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.189677 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.189689 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.214988 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.232321 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.240263 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-config-data" (OuterVolumeSpecName: "config-data") pod "4b795408-0144-4d73-8a56-043ce3e38ee6" (UID: "4b795408-0144-4d73-8a56-043ce3e38ee6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.245719 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "ae4b38bb-7bda-45e9-9d87-f79fe72e9008" (UID: "ae4b38bb-7bda-45e9-9d87-f79fe72e9008"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.247296 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41309df4-53d7-42a8-8a5b-4816cc25fdfb" (UID: "41309df4-53d7-42a8-8a5b-4816cc25fdfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.291934 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.291978 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.291991 4770 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae4b38bb-7bda-45e9-9d87-f79fe72e9008-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.299899 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-config-data" (OuterVolumeSpecName: "config-data") pod "79ab7038-39be-4c09-82ca-00f83aa69e37" (UID: "79ab7038-39be-4c09-82ca-00f83aa69e37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.320300 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-config-data" (OuterVolumeSpecName: "config-data") pod "d49edb80-7b9b-4b10-afbc-652ec540542e" (UID: "d49edb80-7b9b-4b10-afbc-652ec540542e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.357524 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4b795408-0144-4d73-8a56-043ce3e38ee6" (UID: "4b795408-0144-4d73-8a56-043ce3e38ee6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.380606 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "79ab7038-39be-4c09-82ca-00f83aa69e37" (UID: "79ab7038-39be-4c09-82ca-00f83aa69e37"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.389893 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3321-account-delete-gmw9l" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.400341 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "83d745cd-57f5-4624-9d2c-75353e52fa56" (UID: "83d745cd-57f5-4624-9d2c-75353e52fa56"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.400413 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4b795408-0144-4d73-8a56-043ce3e38ee6" (UID: "4b795408-0144-4d73-8a56-043ce3e38ee6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.405296 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b795408-0144-4d73-8a56-043ce3e38ee6" (UID: "4b795408-0144-4d73-8a56-043ce3e38ee6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.429109 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.429150 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.429162 4770 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.429173 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79ab7038-39be-4c09-82ca-00f83aa69e37-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.429193 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b795408-0144-4d73-8a56-043ce3e38ee6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.429205 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.429218 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d49edb80-7b9b-4b10-afbc-652ec540542e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.452411 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "83d745cd-57f5-4624-9d2c-75353e52fa56" (UID: "83d745cd-57f5-4624-9d2c-75353e52fa56"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.452609 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data" (OuterVolumeSpecName: "config-data") pod "83d745cd-57f5-4624-9d2c-75353e52fa56" (UID: "83d745cd-57f5-4624-9d2c-75353e52fa56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.453720 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83d745cd-57f5-4624-9d2c-75353e52fa56" (UID: "83d745cd-57f5-4624-9d2c-75353e52fa56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.460544 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "41309df4-53d7-42a8-8a5b-4816cc25fdfb" (UID: "41309df4-53d7-42a8-8a5b-4816cc25fdfb"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.490493 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "76f9865e-be12-4821-a088-956061126f7b" (UID: "76f9865e-be12-4821-a088-956061126f7b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.498631 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "76f9865e-be12-4821-a088-956061126f7b" (UID: "76f9865e-be12-4821-a088-956061126f7b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.531021 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbttb\" (UniqueName: \"kubernetes.io/projected/6d5328f9-e5e9-4ad5-96ff-02442117ff0e-kube-api-access-jbttb\") pod \"6d5328f9-e5e9-4ad5-96ff-02442117ff0e\" (UID: \"6d5328f9-e5e9-4ad5-96ff-02442117ff0e\") " Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.531728 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.531833 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.531888 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.531962 4770 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/41309df4-53d7-42a8-8a5b-4816cc25fdfb-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.532036 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d745cd-57f5-4624-9d2c-75353e52fa56-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.532093 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/76f9865e-be12-4821-a088-956061126f7b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.534129 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d5328f9-e5e9-4ad5-96ff-02442117ff0e-kube-api-access-jbttb" (OuterVolumeSpecName: "kube-api-access-jbttb") pod "6d5328f9-e5e9-4ad5-96ff-02442117ff0e" (UID: "6d5328f9-e5e9-4ad5-96ff-02442117ff0e"). InnerVolumeSpecName "kube-api-access-jbttb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.567283 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3321-account-delete-gmw9l" event={"ID":"6d5328f9-e5e9-4ad5-96ff-02442117ff0e","Type":"ContainerDied","Data":"e3c582c0354b608d8eac07ebe1b7bb0a1cf369dc801f51dac96dcff1d7f893d3"} Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.567178 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3321-account-delete-gmw9l" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.567774 4770 scope.go:117] "RemoveContainer" containerID="2eb35e9ba346c7f87efe9e066071c98d4374aee1f72cd61b4a26a7bd7f997002" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.584289 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystonef041-account-delete-d8qs9" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.586485 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.587837 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.589305 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dbfd75b46-z42g2" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.590098 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.590834 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.591330 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb0c9-account-delete-hc97s" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.592449 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f85c64c7b-jcqvq" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.594029 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.594172 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.594761 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"41309df4-53d7-42a8-8a5b-4816cc25fdfb","Type":"ContainerDied","Data":"e43aa5b60dcc68b523a2831b338d417300dba571425330b9d39c087939b534fe"} Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.623959 4770 scope.go:117] "RemoveContainer" containerID="5005796694c28178137720d1fa9150a9692d20f97c8d47bfb7d90339bebdf44b" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.637825 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbttb\" (UniqueName: \"kubernetes.io/projected/6d5328f9-e5e9-4ad5-96ff-02442117ff0e-kube-api-access-jbttb\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.745721 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06da071b-e1a1-42da-9e6e-fd957923876d" path="/var/lib/kubelet/pods/06da071b-e1a1-42da-9e6e-fd957923876d/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.749501 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d236ba3-55db-460a-9daa-d6e6e8b06766" path="/var/lib/kubelet/pods/1d236ba3-55db-460a-9daa-d6e6e8b06766/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.750121 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2af9166c-e2c5-462d-be52-340be8aadfd5" path="/var/lib/kubelet/pods/2af9166c-e2c5-462d-be52-340be8aadfd5/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.750814 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35f43e87-e3d9-4508-8bb1-4fe89c64017f" path="/var/lib/kubelet/pods/35f43e87-e3d9-4508-8bb1-4fe89c64017f/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.751880 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c47d914-960a-4c5e-a6f3-584998910e09" path="/var/lib/kubelet/pods/4c47d914-960a-4c5e-a6f3-584998910e09/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.752574 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="697bd5d5-22c6-4303-9077-434eaa099f4a" path="/var/lib/kubelet/pods/697bd5d5-22c6-4303-9077-434eaa099f4a/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.753290 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" path="/var/lib/kubelet/pods/87293d1f-5e33-4a9a-9978-67487fd14809/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.754455 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d18c6351-6663-4a3c-84b6-2752f5e66e49" path="/var/lib/kubelet/pods/d18c6351-6663-4a3c-84b6-2752f5e66e49/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.754932 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c" path="/var/lib/kubelet/pods/d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.755585 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" path="/var/lib/kubelet/pods/e328a4dc-7a4d-4cb0-82a8-f60df78328d8/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.756787 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e826d9bb-183e-401b-bc8b-6cfe0938a7f6" path="/var/lib/kubelet/pods/e826d9bb-183e-401b-bc8b-6cfe0938a7f6/volumes" Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.758777 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi3321-account-delete-gmw9l"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.758896 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi3321-account-delete-gmw9l"] Oct 04 03:25:39 crc kubenswrapper[4770]: E1004 03:25:39.845050 4770 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:39 crc kubenswrapper[4770]: E1004 03:25:39.845150 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data podName:642fe06d-e5b8-461e-a1f7-dbcefc071945 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:47.845125327 +0000 UTC m=+1359.137135049 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data") pod "rabbitmq-cell1-server-0" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945") : configmap "rabbitmq-cell1-config-data" not found Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.954299 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronb0c9-account-delete-hc97s"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.977889 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutronb0c9-account-delete-hc97s"] Oct 04 03:25:39 crc kubenswrapper[4770]: I1004 03:25:39.995134 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystonef041-account-delete-d8qs9"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.000589 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystonef041-account-delete-d8qs9"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.006646 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.012207 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.017994 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.023863 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.058223 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.065878 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: E1004 03:25:40.101256 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 03:25:40 crc kubenswrapper[4770]: E1004 03:25:40.105264 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.109198 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f85c64c7b-jcqvq"] Oct 04 03:25:40 crc kubenswrapper[4770]: E1004 03:25:40.109842 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 03:25:40 crc kubenswrapper[4770]: E1004 03:25:40.109896 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="41a2c8f7-f9d1-44bf-8381-ceea9504a909" containerName="nova-scheduler-scheduler" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.158366 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5hjg\" (UniqueName: \"kubernetes.io/projected/e13efcbd-d9b2-4c14-b681-471e7639d401-kube-api-access-h5hjg\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.161098 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6f85c64c7b-jcqvq"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.167780 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.174847 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.183431 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.183826 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.191275 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.196312 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-dbfd75b46-z42g2"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.200321 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-dbfd75b46-z42g2"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.219106 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.232034 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.367061 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-default\") pod \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.367134 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtg8h\" (UniqueName: \"kubernetes.io/projected/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kube-api-access-xtg8h\") pod \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.367190 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-operator-scripts\") pod \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.367243 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-secrets\") pod \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.367262 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-generated\") pod \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.367280 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-galera-tls-certs\") pod \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.367298 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-combined-ca-bundle\") pod \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.367389 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kolla-config\") pod \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.367425 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\" (UID: \"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.368535 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" (UID: "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.368652 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" (UID: "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.368722 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" (UID: "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.369287 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" (UID: "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.388825 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kube-api-access-xtg8h" (OuterVolumeSpecName: "kube-api-access-xtg8h") pod "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" (UID: "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7"). InnerVolumeSpecName "kube-api-access-xtg8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.398090 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "mysql-db") pod "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" (UID: "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.410836 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-secrets" (OuterVolumeSpecName: "secrets") pod "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" (UID: "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.438101 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" (UID: "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.457684 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" (UID: "1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.469288 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.469319 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.469331 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtg8h\" (UniqueName: \"kubernetes.io/projected/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kube-api-access-xtg8h\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.469340 4770 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.469350 4770 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.469359 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.469367 4770 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.469375 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.469384 4770 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.485904 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.572941 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.609284 4770 generic.go:334] "Generic (PLEG): container finished" podID="2def82a8-5927-4ebe-ac87-e8ad232797ee" containerID="19babfd2683be969879e3923b984d72aa3faecdef7166dfa931d52c111a1f74f" exitCode=0 Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.609377 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2def82a8-5927-4ebe-ac87-e8ad232797ee","Type":"ContainerDied","Data":"19babfd2683be969879e3923b984d72aa3faecdef7166dfa931d52c111a1f74f"} Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.627233 4770 generic.go:334] "Generic (PLEG): container finished" podID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" containerID="6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3" exitCode=0 Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.627353 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7","Type":"ContainerDied","Data":"6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3"} Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.627358 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.627397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7","Type":"ContainerDied","Data":"72555b469a440d5f31b527b5d35fb623f29918672fdfd1486433293d3dbd19c1"} Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.627424 4770 scope.go:117] "RemoveContainer" containerID="6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.630321 4770 generic.go:334] "Generic (PLEG): container finished" podID="ee078c16-5858-4c24-b936-11d4aa568f65" containerID="1f5f2c63939ced31d1a6828f3ba1395e0cbc60c93106d579d81b4050e2c43e88" exitCode=0 Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.630391 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-866cc6bf5d-h867j" event={"ID":"ee078c16-5858-4c24-b936-11d4aa568f65","Type":"ContainerDied","Data":"1f5f2c63939ced31d1a6828f3ba1395e0cbc60c93106d579d81b4050e2c43e88"} Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.706154 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.714617 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.729131 4770 scope.go:117] "RemoveContainer" containerID="6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.765163 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.831082 4770 scope.go:117] "RemoveContainer" containerID="6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3" Oct 04 03:25:40 crc kubenswrapper[4770]: E1004 03:25:40.831611 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3\": container with ID starting with 6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3 not found: ID does not exist" containerID="6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.831672 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3"} err="failed to get container status \"6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3\": rpc error: code = NotFound desc = could not find container \"6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3\": container with ID starting with 6dfa5f9bef165a97382a82b4cc7b07a70fcefe2cc68715f94162b3af3f67f9e3 not found: ID does not exist" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.831710 4770 scope.go:117] "RemoveContainer" containerID="6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d" Oct 04 03:25:40 crc kubenswrapper[4770]: E1004 03:25:40.832077 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d\": container with ID starting with 6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d not found: ID does not exist" containerID="6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.832105 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d"} err="failed to get container status \"6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d\": rpc error: code = NotFound desc = could not find container \"6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d\": container with ID starting with 6a0a3dea122b4a92542a73e3decc6a50f651f5355069a3c39730e4f889f8149d not found: ID does not exist" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.882587 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2def82a8-5927-4ebe-ac87-e8ad232797ee-erlang-cookie-secret\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.882676 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-tls\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.882710 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtcmp\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-kube-api-access-xtcmp\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.882746 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-plugins-conf\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.882770 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-confd\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.882811 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.882907 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2def82a8-5927-4ebe-ac87-e8ad232797ee-pod-info\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.882945 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-plugins\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.882980 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-server-conf\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.883032 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-erlang-cookie\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.883052 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"2def82a8-5927-4ebe-ac87-e8ad232797ee\" (UID: \"2def82a8-5927-4ebe-ac87-e8ad232797ee\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.883560 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.884933 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.885385 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.885865 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.887984 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2def82a8-5927-4ebe-ac87-e8ad232797ee-pod-info" (OuterVolumeSpecName: "pod-info") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.888124 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-kube-api-access-xtcmp" (OuterVolumeSpecName: "kube-api-access-xtcmp") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "kube-api-access-xtcmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.891456 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.891548 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2def82a8-5927-4ebe-ac87-e8ad232797ee-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.892523 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.931594 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-server-conf" (OuterVolumeSpecName: "server-conf") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.937658 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data" (OuterVolumeSpecName: "config-data") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.984923 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-internal-tls-certs\") pod \"ee078c16-5858-4c24-b936-11d4aa568f65\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.984973 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-scripts\") pod \"ee078c16-5858-4c24-b936-11d4aa568f65\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985035 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-credential-keys\") pod \"ee078c16-5858-4c24-b936-11d4aa568f65\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985100 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-combined-ca-bundle\") pod \"ee078c16-5858-4c24-b936-11d4aa568f65\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985171 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-config-data\") pod \"ee078c16-5858-4c24-b936-11d4aa568f65\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985312 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcv9d\" (UniqueName: \"kubernetes.io/projected/ee078c16-5858-4c24-b936-11d4aa568f65-kube-api-access-hcv9d\") pod \"ee078c16-5858-4c24-b936-11d4aa568f65\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985350 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-public-tls-certs\") pod \"ee078c16-5858-4c24-b936-11d4aa568f65\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985457 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-fernet-keys\") pod \"ee078c16-5858-4c24-b936-11d4aa568f65\" (UID: \"ee078c16-5858-4c24-b936-11d4aa568f65\") " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985877 4770 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2def82a8-5927-4ebe-ac87-e8ad232797ee-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985896 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985905 4770 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985914 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985936 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985946 4770 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2def82a8-5927-4ebe-ac87-e8ad232797ee-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985956 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985967 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtcmp\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-kube-api-access-xtcmp\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985976 4770 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:40 crc kubenswrapper[4770]: I1004 03:25:40.985984 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2def82a8-5927-4ebe-ac87-e8ad232797ee-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.001670 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee078c16-5858-4c24-b936-11d4aa568f65-kube-api-access-hcv9d" (OuterVolumeSpecName: "kube-api-access-hcv9d") pod "ee078c16-5858-4c24-b936-11d4aa568f65" (UID: "ee078c16-5858-4c24-b936-11d4aa568f65"). InnerVolumeSpecName "kube-api-access-hcv9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.030623 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ee078c16-5858-4c24-b936-11d4aa568f65" (UID: "ee078c16-5858-4c24-b936-11d4aa568f65"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.033157 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-scripts" (OuterVolumeSpecName: "scripts") pod "ee078c16-5858-4c24-b936-11d4aa568f65" (UID: "ee078c16-5858-4c24-b936-11d4aa568f65"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.033750 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ee078c16-5858-4c24-b936-11d4aa568f65" (UID: "ee078c16-5858-4c24-b936-11d4aa568f65"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.067328 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.089184 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee078c16-5858-4c24-b936-11d4aa568f65" (UID: "ee078c16-5858-4c24-b936-11d4aa568f65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.089438 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2def82a8-5927-4ebe-ac87-e8ad232797ee" (UID: "2def82a8-5927-4ebe-ac87-e8ad232797ee"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.096110 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.096143 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2def82a8-5927-4ebe-ac87-e8ad232797ee-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.096154 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcv9d\" (UniqueName: \"kubernetes.io/projected/ee078c16-5858-4c24-b936-11d4aa568f65-kube-api-access-hcv9d\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.096189 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.096201 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.096209 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.096220 4770 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.098886 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-config-data" (OuterVolumeSpecName: "config-data") pod "ee078c16-5858-4c24-b936-11d4aa568f65" (UID: "ee078c16-5858-4c24-b936-11d4aa568f65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.110588 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ee078c16-5858-4c24-b936-11d4aa568f65" (UID: "ee078c16-5858-4c24-b936-11d4aa568f65"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.120863 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ee078c16-5858-4c24-b936-11d4aa568f65" (UID: "ee078c16-5858-4c24-b936-11d4aa568f65"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.198046 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.198080 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.198090 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee078c16-5858-4c24-b936-11d4aa568f65-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.233902 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a36939a0-d10b-488e-8bd4-165b1f9ae4e9/ovn-northd/0.log" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.233993 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.245358 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:25:41 crc kubenswrapper[4770]: E1004 03:25:41.249848 4770 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 04 03:25:41 crc kubenswrapper[4770]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-04T03:25:33Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 04 03:25:41 crc kubenswrapper[4770]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 04 03:25:41 crc kubenswrapper[4770]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-c2v9x" message=< Oct 04 03:25:41 crc kubenswrapper[4770]: Exiting ovn-controller (1) [FAILED] Oct 04 03:25:41 crc kubenswrapper[4770]: Killing ovn-controller (1) [ OK ] Oct 04 03:25:41 crc kubenswrapper[4770]: Killing ovn-controller (1) with SIGKILL [ OK ] Oct 04 03:25:41 crc kubenswrapper[4770]: 2025-10-04T03:25:33Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 04 03:25:41 crc kubenswrapper[4770]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 04 03:25:41 crc kubenswrapper[4770]: > Oct 04 03:25:41 crc kubenswrapper[4770]: E1004 03:25:41.249881 4770 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 04 03:25:41 crc kubenswrapper[4770]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-04T03:25:33Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 04 03:25:41 crc kubenswrapper[4770]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 04 03:25:41 crc kubenswrapper[4770]: > pod="openstack/ovn-controller-c2v9x" podUID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" containerName="ovn-controller" containerID="cri-o://37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.249919 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-c2v9x" podUID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" containerName="ovn-controller" containerID="cri-o://37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30" gracePeriod=21 Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.400936 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph4mp\" (UniqueName: \"kubernetes.io/projected/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-kube-api-access-ph4mp\") pod \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.400981 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401026 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-metrics-certs-tls-certs\") pod \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401048 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-rundir\") pod \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401100 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gcjw\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-kube-api-access-4gcjw\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401126 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-northd-tls-certs\") pod \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401178 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-config\") pod \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401227 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/642fe06d-e5b8-461e-a1f7-dbcefc071945-pod-info\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401244 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-confd\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401281 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-server-conf\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401302 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401334 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-erlang-cookie\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401351 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-combined-ca-bundle\") pod \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401370 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-plugins-conf\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401389 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-scripts\") pod \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\" (UID: \"a36939a0-d10b-488e-8bd4-165b1f9ae4e9\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401416 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-plugins\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401443 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-tls\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.401469 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/642fe06d-e5b8-461e-a1f7-dbcefc071945-erlang-cookie-secret\") pod \"642fe06d-e5b8-461e-a1f7-dbcefc071945\" (UID: \"642fe06d-e5b8-461e-a1f7-dbcefc071945\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.402281 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "a36939a0-d10b-488e-8bd4-165b1f9ae4e9" (UID: "a36939a0-d10b-488e-8bd4-165b1f9ae4e9"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.402483 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.402626 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.402830 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.403054 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-scripts" (OuterVolumeSpecName: "scripts") pod "a36939a0-d10b-488e-8bd4-165b1f9ae4e9" (UID: "a36939a0-d10b-488e-8bd4-165b1f9ae4e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.403958 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-config" (OuterVolumeSpecName: "config") pod "a36939a0-d10b-488e-8bd4-165b1f9ae4e9" (UID: "a36939a0-d10b-488e-8bd4-165b1f9ae4e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.404413 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.412118 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-kube-api-access-4gcjw" (OuterVolumeSpecName: "kube-api-access-4gcjw") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "kube-api-access-4gcjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.412634 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.412795 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-kube-api-access-ph4mp" (OuterVolumeSpecName: "kube-api-access-ph4mp") pod "a36939a0-d10b-488e-8bd4-165b1f9ae4e9" (UID: "a36939a0-d10b-488e-8bd4-165b1f9ae4e9"). InnerVolumeSpecName "kube-api-access-ph4mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.416596 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642fe06d-e5b8-461e-a1f7-dbcefc071945-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.423165 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/642fe06d-e5b8-461e-a1f7-dbcefc071945-pod-info" (OuterVolumeSpecName: "pod-info") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.427540 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data" (OuterVolumeSpecName: "config-data") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.438745 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a36939a0-d10b-488e-8bd4-165b1f9ae4e9" (UID: "a36939a0-d10b-488e-8bd4-165b1f9ae4e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.487629 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-server-conf" (OuterVolumeSpecName: "server-conf") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.489583 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "a36939a0-d10b-488e-8bd4-165b1f9ae4e9" (UID: "a36939a0-d10b-488e-8bd4-165b1f9ae4e9"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.489685 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a36939a0-d10b-488e-8bd4-165b1f9ae4e9" (UID: "a36939a0-d10b-488e-8bd4-165b1f9ae4e9"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503797 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503829 4770 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/642fe06d-e5b8-461e-a1f7-dbcefc071945-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503842 4770 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503851 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503862 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503874 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503882 4770 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/642fe06d-e5b8-461e-a1f7-dbcefc071945-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503891 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503899 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503907 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503915 4770 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/642fe06d-e5b8-461e-a1f7-dbcefc071945-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503923 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph4mp\" (UniqueName: \"kubernetes.io/projected/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-kube-api-access-ph4mp\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.503953 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.504027 4770 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.504042 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.504051 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gcjw\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-kube-api-access-4gcjw\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.504059 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a36939a0-d10b-488e-8bd4-165b1f9ae4e9-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.524331 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.561498 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "642fe06d-e5b8-461e-a1f7-dbcefc071945" (UID: "642fe06d-e5b8-461e-a1f7-dbcefc071945"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.606297 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.606355 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/642fe06d-e5b8-461e-a1f7-dbcefc071945-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.630063 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-c2v9x_7916b1d6-76ad-4780-bfc2-2a0a738b8b8d/ovn-controller/0.log" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.630139 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.692585 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" path="/var/lib/kubelet/pods/1cb81588-b9f4-4d8a-8baf-33aebaa6e64d/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.693331 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" path="/var/lib/kubelet/pods/1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.693935 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41309df4-53d7-42a8-8a5b-4816cc25fdfb" path="/var/lib/kubelet/pods/41309df4-53d7-42a8-8a5b-4816cc25fdfb/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.704641 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" path="/var/lib/kubelet/pods/4b795408-0144-4d73-8a56-043ce3e38ee6/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.705155 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d5328f9-e5e9-4ad5-96ff-02442117ff0e" path="/var/lib/kubelet/pods/6d5328f9-e5e9-4ad5-96ff-02442117ff0e/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.705643 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76f9865e-be12-4821-a088-956061126f7b" path="/var/lib/kubelet/pods/76f9865e-be12-4821-a088-956061126f7b/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.709091 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" path="/var/lib/kubelet/pods/79ab7038-39be-4c09-82ca-00f83aa69e37/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.709633 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" path="/var/lib/kubelet/pods/83d745cd-57f5-4624-9d2c-75353e52fa56/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.712986 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae4b38bb-7bda-45e9-9d87-f79fe72e9008" path="/var/lib/kubelet/pods/ae4b38bb-7bda-45e9-9d87-f79fe72e9008/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.713436 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c00a8476-9758-46f5-864f-9a7b80ca8bd1" path="/var/lib/kubelet/pods/c00a8476-9758-46f5-864f-9a7b80ca8bd1/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.713946 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d49edb80-7b9b-4b10-afbc-652ec540542e" path="/var/lib/kubelet/pods/d49edb80-7b9b-4b10-afbc-652ec540542e/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.714513 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e13efcbd-d9b2-4c14-b681-471e7639d401" path="/var/lib/kubelet/pods/e13efcbd-d9b2-4c14-b681-471e7639d401/volumes" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.716598 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.719619 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2def82a8-5927-4ebe-ac87-e8ad232797ee","Type":"ContainerDied","Data":"b7998530f1dfa56d5f6be2c686686a50395bb8d934c54b5975ea3d450814dcf5"} Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.719667 4770 scope.go:117] "RemoveContainer" containerID="19babfd2683be969879e3923b984d72aa3faecdef7166dfa931d52c111a1f74f" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.733476 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-c2v9x_7916b1d6-76ad-4780-bfc2-2a0a738b8b8d/ovn-controller/0.log" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.733517 4770 generic.go:334] "Generic (PLEG): container finished" podID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" containerID="37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30" exitCode=137 Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.733569 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x" event={"ID":"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d","Type":"ContainerDied","Data":"37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30"} Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.733596 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-c2v9x" event={"ID":"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d","Type":"ContainerDied","Data":"6d25180e1bdf6781040961e6aad0cc608329ce28985693de6216fcd4475a20cc"} Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.733642 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-c2v9x" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.746492 4770 generic.go:334] "Generic (PLEG): container finished" podID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerID="6fd96f4fbeb0043266b189a92e71e234f2a99b6b94cea0a2db77c70f4d686a5b" exitCode=0 Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.746558 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerDied","Data":"6fd96f4fbeb0043266b189a92e71e234f2a99b6b94cea0a2db77c70f4d686a5b"} Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.755362 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-866cc6bf5d-h867j" event={"ID":"ee078c16-5858-4c24-b936-11d4aa568f65","Type":"ContainerDied","Data":"96e2dfbcc9586566650da34c2453a3a3a0c2db26296fc47219d1c0dad11311c4"} Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.755445 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-866cc6bf5d-h867j" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.758451 4770 generic.go:334] "Generic (PLEG): container finished" podID="642fe06d-e5b8-461e-a1f7-dbcefc071945" containerID="4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749" exitCode=0 Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.758593 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"642fe06d-e5b8-461e-a1f7-dbcefc071945","Type":"ContainerDied","Data":"4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749"} Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.758647 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"642fe06d-e5b8-461e-a1f7-dbcefc071945","Type":"ContainerDied","Data":"e3aab1fd6e666025903602376b35fc475759dabf57ddeba8d7c0b7252952aa50"} Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.758561 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.762834 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a36939a0-d10b-488e-8bd4-165b1f9ae4e9/ovn-northd/0.log" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.762898 4770 generic.go:334] "Generic (PLEG): container finished" podID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" exitCode=139 Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.762932 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a36939a0-d10b-488e-8bd4-165b1f9ae4e9","Type":"ContainerDied","Data":"38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed"} Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.762962 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a36939a0-d10b-488e-8bd4-165b1f9ae4e9","Type":"ContainerDied","Data":"32dcbadbecab7ac088279e37bd475ff16f530915676c670fa5c5005ecd755ba0"} Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.763054 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.813644 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run-ovn\") pod \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.813757 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-combined-ca-bundle\") pod \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.813805 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-log-ovn\") pod \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.813827 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run\") pod \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.813850 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-ovn-controller-tls-certs\") pod \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.813862 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" (UID: "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.813929 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-scripts\") pod \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.813946 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" (UID: "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.813957 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58bkz\" (UniqueName: \"kubernetes.io/projected/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-kube-api-access-58bkz\") pod \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\" (UID: \"7916b1d6-76ad-4780-bfc2-2a0a738b8b8d\") " Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.814499 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run" (OuterVolumeSpecName: "var-run") pod "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" (UID: "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.814980 4770 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.815044 4770 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.815062 4770 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.817781 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-scripts" (OuterVolumeSpecName: "scripts") pod "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" (UID: "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.819494 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-kube-api-access-58bkz" (OuterVolumeSpecName: "kube-api-access-58bkz") pod "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" (UID: "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d"). InnerVolumeSpecName "kube-api-access-58bkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.864293 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" (UID: "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.872867 4770 scope.go:117] "RemoveContainer" containerID="5305bb132085306589edf135617a3aa23523ea917b7840da309c820a5a534359" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.922583 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.922645 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.922662 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58bkz\" (UniqueName: \"kubernetes.io/projected/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-kube-api-access-58bkz\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.936376 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-866cc6bf5d-h867j"] Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.941387 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-866cc6bf5d-h867j"] Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.955686 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.967079 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:25:41 crc kubenswrapper[4770]: I1004 03:25:41.969471 4770 scope.go:117] "RemoveContainer" containerID="37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.010388 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" (UID: "7916b1d6-76ad-4780-bfc2-2a0a738b8b8d"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.038724 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.093563 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.108952 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.121740 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.156268 4770 scope.go:117] "RemoveContainer" containerID="37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30" Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.159527 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30\": container with ID starting with 37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30 not found: ID does not exist" containerID="37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.159574 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30"} err="failed to get container status \"37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30\": rpc error: code = NotFound desc = could not find container \"37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30\": container with ID starting with 37e0d43cdd1a9353dabdc29854112d6411ca168e3315cdd14966dab3f67b9c30 not found: ID does not exist" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.159605 4770 scope.go:117] "RemoveContainer" containerID="1f5f2c63939ced31d1a6828f3ba1395e0cbc60c93106d579d81b4050e2c43e88" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.162221 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.175083 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.194462 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-c2v9x"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.195062 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-ceilometer-tls-certs\") pod \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.195099 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzp9q\" (UniqueName: \"kubernetes.io/projected/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-kube-api-access-fzp9q\") pod \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.195137 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-config-data\") pod \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.195225 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-run-httpd\") pod \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.195289 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-scripts\") pod \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.195318 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-combined-ca-bundle\") pod \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.195347 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-log-httpd\") pod \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.195429 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-sg-core-conf-yaml\") pod \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\" (UID: \"0d4f43f8-c9a5-4a80-aab0-2b9226499b27\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.198554 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0d4f43f8-c9a5-4a80-aab0-2b9226499b27" (UID: "0d4f43f8-c9a5-4a80-aab0-2b9226499b27"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.198789 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0d4f43f8-c9a5-4a80-aab0-2b9226499b27" (UID: "0d4f43f8-c9a5-4a80-aab0-2b9226499b27"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.198952 4770 scope.go:117] "RemoveContainer" containerID="4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.204293 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-kube-api-access-fzp9q" (OuterVolumeSpecName: "kube-api-access-fzp9q") pod "0d4f43f8-c9a5-4a80-aab0-2b9226499b27" (UID: "0d4f43f8-c9a5-4a80-aab0-2b9226499b27"). InnerVolumeSpecName "kube-api-access-fzp9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.205293 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-scripts" (OuterVolumeSpecName: "scripts") pod "0d4f43f8-c9a5-4a80-aab0-2b9226499b27" (UID: "0d4f43f8-c9a5-4a80-aab0-2b9226499b27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.207117 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-c2v9x"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.230082 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0d4f43f8-c9a5-4a80-aab0-2b9226499b27" (UID: "0d4f43f8-c9a5-4a80-aab0-2b9226499b27"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.252359 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0d4f43f8-c9a5-4a80-aab0-2b9226499b27" (UID: "0d4f43f8-c9a5-4a80-aab0-2b9226499b27"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.274886 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d4f43f8-c9a5-4a80-aab0-2b9226499b27" (UID: "0d4f43f8-c9a5-4a80-aab0-2b9226499b27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.297454 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.297718 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.297801 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.297858 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.297935 4770 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.297998 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzp9q\" (UniqueName: \"kubernetes.io/projected/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-kube-api-access-fzp9q\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.298099 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.297541 4770 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.298245 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:50.298228013 +0000 UTC m=+1361.590237725 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-worker-config-data" not found Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.297644 4770 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.298375 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:25:50.298367956 +0000 UTC m=+1361.590377668 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-config-data" not found Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.300049 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-config-data" (OuterVolumeSpecName: "config-data") pod "0d4f43f8-c9a5-4a80-aab0-2b9226499b27" (UID: "0d4f43f8-c9a5-4a80-aab0-2b9226499b27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.323959 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.335050 4770 scope.go:117] "RemoveContainer" containerID="0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.399600 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-combined-ca-bundle\") pod \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.399675 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-config-data\") pod \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.399717 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngntj\" (UniqueName: \"kubernetes.io/projected/41a2c8f7-f9d1-44bf-8381-ceea9504a909-kube-api-access-ngntj\") pod \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\" (UID: \"41a2c8f7-f9d1-44bf-8381-ceea9504a909\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.400179 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4f43f8-c9a5-4a80-aab0-2b9226499b27-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.403537 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41a2c8f7-f9d1-44bf-8381-ceea9504a909-kube-api-access-ngntj" (OuterVolumeSpecName: "kube-api-access-ngntj") pod "41a2c8f7-f9d1-44bf-8381-ceea9504a909" (UID: "41a2c8f7-f9d1-44bf-8381-ceea9504a909"). InnerVolumeSpecName "kube-api-access-ngntj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.425592 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-config-data" (OuterVolumeSpecName: "config-data") pod "41a2c8f7-f9d1-44bf-8381-ceea9504a909" (UID: "41a2c8f7-f9d1-44bf-8381-ceea9504a909"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.427776 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41a2c8f7-f9d1-44bf-8381-ceea9504a909" (UID: "41a2c8f7-f9d1-44bf-8381-ceea9504a909"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.436900 4770 scope.go:117] "RemoveContainer" containerID="4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749" Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.437433 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749\": container with ID starting with 4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749 not found: ID does not exist" containerID="4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.437475 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749"} err="failed to get container status \"4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749\": rpc error: code = NotFound desc = could not find container \"4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749\": container with ID starting with 4807552ad4f1914decdaa369d8abf729e02bedb641580f5c94b31cfc9338b749 not found: ID does not exist" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.437505 4770 scope.go:117] "RemoveContainer" containerID="0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01" Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.437878 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01\": container with ID starting with 0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01 not found: ID does not exist" containerID="0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.437906 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01"} err="failed to get container status \"0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01\": rpc error: code = NotFound desc = could not find container \"0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01\": container with ID starting with 0fa90352ee06f295de6457b6607bcaed5514c5ac813f2f356b4ac2a715df1a01 not found: ID does not exist" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.437925 4770 scope.go:117] "RemoveContainer" containerID="00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.461548 4770 scope.go:117] "RemoveContainer" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.482989 4770 scope.go:117] "RemoveContainer" containerID="00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90" Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.483373 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90\": container with ID starting with 00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90 not found: ID does not exist" containerID="00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.483419 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90"} err="failed to get container status \"00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90\": rpc error: code = NotFound desc = could not find container \"00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90\": container with ID starting with 00b2b3662b422dcf066422de937fe1413d00f26c77bb14eeeb569b022c337e90 not found: ID does not exist" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.483441 4770 scope.go:117] "RemoveContainer" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.484169 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed\": container with ID starting with 38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed not found: ID does not exist" containerID="38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.484234 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed"} err="failed to get container status \"38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed\": rpc error: code = NotFound desc = could not find container \"38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed\": container with ID starting with 38c4fb3775fdd9a3625906c7e50f2d480fae4decec51aefc09d3c6b920c20fed not found: ID does not exist" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.502924 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.502968 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41a2c8f7-f9d1-44bf-8381-ceea9504a909-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.502977 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngntj\" (UniqueName: \"kubernetes.io/projected/41a2c8f7-f9d1-44bf-8381-ceea9504a909-kube-api-access-ngntj\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.687551 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.782864 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d4f43f8-c9a5-4a80-aab0-2b9226499b27","Type":"ContainerDied","Data":"cbe9489fa8c679d8b66c4b01c781525a45f303d607c8209d7b432b97f8583077"} Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.783254 4770 scope.go:117] "RemoveContainer" containerID="f6cc4fe8efa2daaa6113859b79bc85ffd8f259a17e76cdbe64107c051e5c6660" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.783480 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.791453 4770 generic.go:334] "Generic (PLEG): container finished" podID="41a2c8f7-f9d1-44bf-8381-ceea9504a909" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" exitCode=0 Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.791496 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.791514 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"41a2c8f7-f9d1-44bf-8381-ceea9504a909","Type":"ContainerDied","Data":"c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa"} Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.791984 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"41a2c8f7-f9d1-44bf-8381-ceea9504a909","Type":"ContainerDied","Data":"769613dedd149600ee44a38e77feeb871c18558cd7951f9032fbd9ff8812e13b"} Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.797040 4770 generic.go:334] "Generic (PLEG): container finished" podID="ef58a910-f85e-4e14-95b5-807741097485" containerID="7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda" exitCode=0 Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.797101 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.797116 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ef58a910-f85e-4e14-95b5-807741097485","Type":"ContainerDied","Data":"7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda"} Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.797160 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ef58a910-f85e-4e14-95b5-807741097485","Type":"ContainerDied","Data":"7078e60ebe9016e8d310ce62e5d638d0400876fabbb7b01b95c7c917d5a6fdf3"} Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.804483 4770 scope.go:117] "RemoveContainer" containerID="3ebcae43ad55173cedbd5766bb36e217f05f019084426c4bf34462bc4520a90f" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.809131 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-combined-ca-bundle\") pod \"ef58a910-f85e-4e14-95b5-807741097485\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.809228 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnfqr\" (UniqueName: \"kubernetes.io/projected/ef58a910-f85e-4e14-95b5-807741097485-kube-api-access-rnfqr\") pod \"ef58a910-f85e-4e14-95b5-807741097485\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.809302 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-config-data\") pod \"ef58a910-f85e-4e14-95b5-807741097485\" (UID: \"ef58a910-f85e-4e14-95b5-807741097485\") " Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.818341 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef58a910-f85e-4e14-95b5-807741097485-kube-api-access-rnfqr" (OuterVolumeSpecName: "kube-api-access-rnfqr") pod "ef58a910-f85e-4e14-95b5-807741097485" (UID: "ef58a910-f85e-4e14-95b5-807741097485"). InnerVolumeSpecName "kube-api-access-rnfqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.837207 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.837337 4770 scope.go:117] "RemoveContainer" containerID="6fd96f4fbeb0043266b189a92e71e234f2a99b6b94cea0a2db77c70f4d686a5b" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.840900 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-config-data" (OuterVolumeSpecName: "config-data") pod "ef58a910-f85e-4e14-95b5-807741097485" (UID: "ef58a910-f85e-4e14-95b5-807741097485"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.842616 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.847189 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef58a910-f85e-4e14-95b5-807741097485" (UID: "ef58a910-f85e-4e14-95b5-807741097485"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.851149 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.856578 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.861616 4770 scope.go:117] "RemoveContainer" containerID="3eef293d1922aeebed6a824d6768b260279f3c458835051702da305ac78ec1e4" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.888842 4770 scope.go:117] "RemoveContainer" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.911287 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.911323 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnfqr\" (UniqueName: \"kubernetes.io/projected/ef58a910-f85e-4e14-95b5-807741097485-kube-api-access-rnfqr\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.911337 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef58a910-f85e-4e14-95b5-807741097485-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.913287 4770 scope.go:117] "RemoveContainer" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.913647 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa\": container with ID starting with c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa not found: ID does not exist" containerID="c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.913695 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa"} err="failed to get container status \"c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa\": rpc error: code = NotFound desc = could not find container \"c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa\": container with ID starting with c82f853fb13f5f9f8793ce4169e4d44d9377af0afecae011a260d1c10173a7fa not found: ID does not exist" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.913721 4770 scope.go:117] "RemoveContainer" containerID="7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.955320 4770 scope.go:117] "RemoveContainer" containerID="7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda" Oct 04 03:25:42 crc kubenswrapper[4770]: E1004 03:25:42.955722 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda\": container with ID starting with 7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda not found: ID does not exist" containerID="7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda" Oct 04 03:25:42 crc kubenswrapper[4770]: I1004 03:25:42.955759 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda"} err="failed to get container status \"7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda\": rpc error: code = NotFound desc = could not find container \"7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda\": container with ID starting with 7f941f9f42dd20d23c2d9b80b47d2b62ca146bae787b4d46b8cc008374caadda not found: ID does not exist" Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.140712 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.146745 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 03:25:43 crc kubenswrapper[4770]: E1004 03:25:43.563352 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:43 crc kubenswrapper[4770]: E1004 03:25:43.563998 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:43 crc kubenswrapper[4770]: E1004 03:25:43.564617 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:43 crc kubenswrapper[4770]: E1004 03:25:43.564670 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" Oct 04 03:25:43 crc kubenswrapper[4770]: E1004 03:25:43.564982 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:43 crc kubenswrapper[4770]: E1004 03:25:43.566214 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:43 crc kubenswrapper[4770]: E1004 03:25:43.567962 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:43 crc kubenswrapper[4770]: E1004 03:25:43.567993 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.684442 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" path="/var/lib/kubelet/pods/0d4f43f8-c9a5-4a80-aab0-2b9226499b27/volumes" Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.686095 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2def82a8-5927-4ebe-ac87-e8ad232797ee" path="/var/lib/kubelet/pods/2def82a8-5927-4ebe-ac87-e8ad232797ee/volumes" Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.687571 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41a2c8f7-f9d1-44bf-8381-ceea9504a909" path="/var/lib/kubelet/pods/41a2c8f7-f9d1-44bf-8381-ceea9504a909/volumes" Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.688675 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="642fe06d-e5b8-461e-a1f7-dbcefc071945" path="/var/lib/kubelet/pods/642fe06d-e5b8-461e-a1f7-dbcefc071945/volumes" Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.689606 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" path="/var/lib/kubelet/pods/7916b1d6-76ad-4780-bfc2-2a0a738b8b8d/volumes" Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.691404 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" path="/var/lib/kubelet/pods/a36939a0-d10b-488e-8bd4-165b1f9ae4e9/volumes" Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.692313 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee078c16-5858-4c24-b936-11d4aa568f65" path="/var/lib/kubelet/pods/ee078c16-5858-4c24-b936-11d4aa568f65/volumes" Oct 04 03:25:43 crc kubenswrapper[4770]: I1004 03:25:43.693113 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef58a910-f85e-4e14-95b5-807741097485" path="/var/lib/kubelet/pods/ef58a910-f85e-4e14-95b5-807741097485/volumes" Oct 04 03:25:48 crc kubenswrapper[4770]: E1004 03:25:48.565201 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:48 crc kubenswrapper[4770]: E1004 03:25:48.567601 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:48 crc kubenswrapper[4770]: E1004 03:25:48.567927 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:48 crc kubenswrapper[4770]: E1004 03:25:48.567974 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" Oct 04 03:25:48 crc kubenswrapper[4770]: E1004 03:25:48.568456 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:48 crc kubenswrapper[4770]: E1004 03:25:48.570197 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:48 crc kubenswrapper[4770]: E1004 03:25:48.571734 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:48 crc kubenswrapper[4770]: E1004 03:25:48.571783 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" Oct 04 03:25:50 crc kubenswrapper[4770]: E1004 03:25:50.338050 4770 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 04 03:25:50 crc kubenswrapper[4770]: E1004 03:25:50.338134 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:06.338115869 +0000 UTC m=+1377.630125581 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-worker-config-data" not found Oct 04 03:25:50 crc kubenswrapper[4770]: E1004 03:25:50.338158 4770 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 04 03:25:50 crc kubenswrapper[4770]: E1004 03:25:50.338251 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:06.338227022 +0000 UTC m=+1377.630236744 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-config-data" not found Oct 04 03:25:53 crc kubenswrapper[4770]: E1004 03:25:53.563335 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:53 crc kubenswrapper[4770]: E1004 03:25:53.565201 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:53 crc kubenswrapper[4770]: E1004 03:25:53.565868 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:53 crc kubenswrapper[4770]: E1004 03:25:53.565959 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" Oct 04 03:25:53 crc kubenswrapper[4770]: E1004 03:25:53.567492 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:53 crc kubenswrapper[4770]: E1004 03:25:53.569687 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:53 crc kubenswrapper[4770]: E1004 03:25:53.571735 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:53 crc kubenswrapper[4770]: E1004 03:25:53.571823 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.898869 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.942727 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-combined-ca-bundle\") pod \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.942795 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csxq4\" (UniqueName: \"kubernetes.io/projected/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-kube-api-access-csxq4\") pod \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.942829 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-ovndb-tls-certs\") pod \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.942855 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-public-tls-certs\") pod \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.942876 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-httpd-config\") pod \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.942997 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-internal-tls-certs\") pod \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.943077 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-config\") pod \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\" (UID: \"145bb94b-ff5d-4b2f-97b1-1b225fc4f164\") " Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.961276 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-kube-api-access-csxq4" (OuterVolumeSpecName: "kube-api-access-csxq4") pod "145bb94b-ff5d-4b2f-97b1-1b225fc4f164" (UID: "145bb94b-ff5d-4b2f-97b1-1b225fc4f164"). InnerVolumeSpecName "kube-api-access-csxq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.961417 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "145bb94b-ff5d-4b2f-97b1-1b225fc4f164" (UID: "145bb94b-ff5d-4b2f-97b1-1b225fc4f164"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.990859 4770 generic.go:334] "Generic (PLEG): container finished" podID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerID="fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85" exitCode=0 Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.990947 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55cdcb545c-rfq2m" Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.990960 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55cdcb545c-rfq2m" event={"ID":"145bb94b-ff5d-4b2f-97b1-1b225fc4f164","Type":"ContainerDied","Data":"fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85"} Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.991718 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55cdcb545c-rfq2m" event={"ID":"145bb94b-ff5d-4b2f-97b1-1b225fc4f164","Type":"ContainerDied","Data":"b4f035e89e59b0a3d2e150ddf9a3f79cd9dce5507edc4a105d60c8dcbcfa2951"} Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.991749 4770 scope.go:117] "RemoveContainer" containerID="886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97" Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.996145 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "145bb94b-ff5d-4b2f-97b1-1b225fc4f164" (UID: "145bb94b-ff5d-4b2f-97b1-1b225fc4f164"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:55 crc kubenswrapper[4770]: I1004 03:25:55.997312 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-config" (OuterVolumeSpecName: "config") pod "145bb94b-ff5d-4b2f-97b1-1b225fc4f164" (UID: "145bb94b-ff5d-4b2f-97b1-1b225fc4f164"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.001722 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "145bb94b-ff5d-4b2f-97b1-1b225fc4f164" (UID: "145bb94b-ff5d-4b2f-97b1-1b225fc4f164"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.013548 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "145bb94b-ff5d-4b2f-97b1-1b225fc4f164" (UID: "145bb94b-ff5d-4b2f-97b1-1b225fc4f164"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.019410 4770 scope.go:117] "RemoveContainer" containerID="fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.027326 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "145bb94b-ff5d-4b2f-97b1-1b225fc4f164" (UID: "145bb94b-ff5d-4b2f-97b1-1b225fc4f164"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.042248 4770 scope.go:117] "RemoveContainer" containerID="886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97" Oct 04 03:25:56 crc kubenswrapper[4770]: E1004 03:25:56.042774 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97\": container with ID starting with 886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97 not found: ID does not exist" containerID="886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.042828 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97"} err="failed to get container status \"886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97\": rpc error: code = NotFound desc = could not find container \"886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97\": container with ID starting with 886ea6e00c8dc57c4cd650e1244542300dacba2ae1a58199e7c399ed30266e97 not found: ID does not exist" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.042867 4770 scope.go:117] "RemoveContainer" containerID="fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85" Oct 04 03:25:56 crc kubenswrapper[4770]: E1004 03:25:56.043420 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85\": container with ID starting with fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85 not found: ID does not exist" containerID="fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.043571 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85"} err="failed to get container status \"fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85\": rpc error: code = NotFound desc = could not find container \"fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85\": container with ID starting with fd50eb62550ec7b4d12d7d4d83d99f29a0200099e94fd7dbf3d5a5b9d42dbb85 not found: ID does not exist" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.044964 4770 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.045109 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.045191 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.045248 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csxq4\" (UniqueName: \"kubernetes.io/projected/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-kube-api-access-csxq4\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.045316 4770 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.045370 4770 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.045421 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145bb94b-ff5d-4b2f-97b1-1b225fc4f164-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.326829 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-55cdcb545c-rfq2m"] Oct 04 03:25:56 crc kubenswrapper[4770]: I1004 03:25:56.335966 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-55cdcb545c-rfq2m"] Oct 04 03:25:57 crc kubenswrapper[4770]: I1004 03:25:57.687606 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" path="/var/lib/kubelet/pods/145bb94b-ff5d-4b2f-97b1-1b225fc4f164/volumes" Oct 04 03:25:58 crc kubenswrapper[4770]: E1004 03:25:58.562601 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:58 crc kubenswrapper[4770]: E1004 03:25:58.563407 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:58 crc kubenswrapper[4770]: E1004 03:25:58.563808 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 04 03:25:58 crc kubenswrapper[4770]: E1004 03:25:58.563908 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" Oct 04 03:25:58 crc kubenswrapper[4770]: E1004 03:25:58.564077 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:58 crc kubenswrapper[4770]: E1004 03:25:58.565408 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:58 crc kubenswrapper[4770]: E1004 03:25:58.566614 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 04 03:25:58 crc kubenswrapper[4770]: E1004 03:25:58.566666 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6nz6q" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" Oct 04 03:26:02 crc kubenswrapper[4770]: I1004 03:26:02.940468 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6nz6q_c5f2e918-133c-440e-ab52-2a389ddc9ede/ovs-vswitchd/0.log" Oct 04 03:26:02 crc kubenswrapper[4770]: I1004 03:26:02.942296 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:26:02 crc kubenswrapper[4770]: I1004 03:26:02.944132 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071411 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data\") pod \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071519 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5f2e918-133c-440e-ab52-2a389ddc9ede-scripts\") pod \"c5f2e918-133c-440e-ab52-2a389ddc9ede\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071561 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-log\") pod \"c5f2e918-133c-440e-ab52-2a389ddc9ede\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071602 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-run\") pod \"c5f2e918-133c-440e-ab52-2a389ddc9ede\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071631 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-combined-ca-bundle\") pod \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071670 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-lib\") pod \"c5f2e918-133c-440e-ab52-2a389ddc9ede\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071689 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-etc-machine-id\") pod \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071676 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-log" (OuterVolumeSpecName: "var-log") pod "c5f2e918-133c-440e-ab52-2a389ddc9ede" (UID: "c5f2e918-133c-440e-ab52-2a389ddc9ede"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071778 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-etc-ovs\") pod \"c5f2e918-133c-440e-ab52-2a389ddc9ede\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071760 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-run" (OuterVolumeSpecName: "var-run") pod "c5f2e918-133c-440e-ab52-2a389ddc9ede" (UID: "c5f2e918-133c-440e-ab52-2a389ddc9ede"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071824 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-lib" (OuterVolumeSpecName: "var-lib") pod "c5f2e918-133c-440e-ab52-2a389ddc9ede" (UID: "c5f2e918-133c-440e-ab52-2a389ddc9ede"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071857 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "c5f2e918-133c-440e-ab52-2a389ddc9ede" (UID: "c5f2e918-133c-440e-ab52-2a389ddc9ede"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071853 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" (UID: "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071819 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77ql8\" (UniqueName: \"kubernetes.io/projected/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-kube-api-access-77ql8\") pod \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.071954 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-scripts\") pod \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.072044 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data-custom\") pod \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\" (UID: \"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.072084 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glpwm\" (UniqueName: \"kubernetes.io/projected/c5f2e918-133c-440e-ab52-2a389ddc9ede-kube-api-access-glpwm\") pod \"c5f2e918-133c-440e-ab52-2a389ddc9ede\" (UID: \"c5f2e918-133c-440e-ab52-2a389ddc9ede\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.072986 4770 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-lib\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.073001 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.073029 4770 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.073039 4770 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-log\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.073049 4770 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5f2e918-133c-440e-ab52-2a389ddc9ede-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.073189 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5f2e918-133c-440e-ab52-2a389ddc9ede-scripts" (OuterVolumeSpecName: "scripts") pod "c5f2e918-133c-440e-ab52-2a389ddc9ede" (UID: "c5f2e918-133c-440e-ab52-2a389ddc9ede"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.080635 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-kube-api-access-77ql8" (OuterVolumeSpecName: "kube-api-access-77ql8") pod "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" (UID: "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2"). InnerVolumeSpecName "kube-api-access-77ql8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.080814 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-scripts" (OuterVolumeSpecName: "scripts") pod "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" (UID: "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.084172 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5f2e918-133c-440e-ab52-2a389ddc9ede-kube-api-access-glpwm" (OuterVolumeSpecName: "kube-api-access-glpwm") pod "c5f2e918-133c-440e-ab52-2a389ddc9ede" (UID: "c5f2e918-133c-440e-ab52-2a389ddc9ede"). InnerVolumeSpecName "kube-api-access-glpwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.084418 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" (UID: "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.091405 4770 generic.go:334] "Generic (PLEG): container finished" podID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerID="24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b" exitCode=137 Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.091506 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2","Type":"ContainerDied","Data":"24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b"} Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.091571 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1908f1eb-8656-43ee-8532-8b9e4ca8e1c2","Type":"ContainerDied","Data":"e2f9ae58de1810066e718f156dc56e2193540ce888fb2110b94141f99335b8ec"} Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.091498 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.092079 4770 scope.go:117] "RemoveContainer" containerID="1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.098637 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6nz6q_c5f2e918-133c-440e-ab52-2a389ddc9ede/ovs-vswitchd/0.log" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.099813 4770 generic.go:334] "Generic (PLEG): container finished" podID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" exitCode=137 Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.099892 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6nz6q" event={"ID":"c5f2e918-133c-440e-ab52-2a389ddc9ede","Type":"ContainerDied","Data":"51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c"} Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.099927 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6nz6q" event={"ID":"c5f2e918-133c-440e-ab52-2a389ddc9ede","Type":"ContainerDied","Data":"7dcea88773c57dcac7ec8f2a5d3895d9156de475966a38c28b29c007a198d6db"} Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.101082 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6nz6q" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.123410 4770 generic.go:334] "Generic (PLEG): container finished" podID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerID="345b8c0092ebcc04825e922384f0c8ee2e508d97ba89f66363c61e92ce3d2670" exitCode=137 Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.123466 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"345b8c0092ebcc04825e922384f0c8ee2e508d97ba89f66363c61e92ce3d2670"} Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.138748 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-6nz6q"] Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.143578 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-6nz6q"] Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.145403 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" (UID: "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.146542 4770 scope.go:117] "RemoveContainer" containerID="24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.172898 4770 scope.go:117] "RemoveContainer" containerID="1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522" Oct 04 03:26:03 crc kubenswrapper[4770]: E1004 03:26:03.173656 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522\": container with ID starting with 1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522 not found: ID does not exist" containerID="1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.173706 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522"} err="failed to get container status \"1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522\": rpc error: code = NotFound desc = could not find container \"1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522\": container with ID starting with 1c16e3997823e51c45054ec8524774cee8b2e3df1dc768d3cbab08980d150522 not found: ID does not exist" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.173738 4770 scope.go:117] "RemoveContainer" containerID="24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b" Oct 04 03:26:03 crc kubenswrapper[4770]: E1004 03:26:03.174122 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b\": container with ID starting with 24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b not found: ID does not exist" containerID="24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.174178 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b"} err="failed to get container status \"24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b\": rpc error: code = NotFound desc = could not find container \"24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b\": container with ID starting with 24dd34811e74c0fe56aa6ee176e9b90bba63ac6a858e773b5ff4ac69459e287b not found: ID does not exist" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.174218 4770 scope.go:117] "RemoveContainer" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.174332 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77ql8\" (UniqueName: \"kubernetes.io/projected/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-kube-api-access-77ql8\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.174369 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.174382 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.174392 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glpwm\" (UniqueName: \"kubernetes.io/projected/c5f2e918-133c-440e-ab52-2a389ddc9ede-kube-api-access-glpwm\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.174402 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5f2e918-133c-440e-ab52-2a389ddc9ede-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.174409 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.198380 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data" (OuterVolumeSpecName: "config-data") pod "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" (UID: "1908f1eb-8656-43ee-8532-8b9e4ca8e1c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.216378 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.218117 4770 scope.go:117] "RemoveContainer" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.239864 4770 scope.go:117] "RemoveContainer" containerID="b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.278255 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") pod \"49d542d9-ee75-4f69-a103-0abb25d8c891\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.278796 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-lock\") pod \"49d542d9-ee75-4f69-a103-0abb25d8c891\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.280914 4770 scope.go:117] "RemoveContainer" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.280933 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-cache\") pod \"49d542d9-ee75-4f69-a103-0abb25d8c891\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.281000 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdfg5\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-kube-api-access-fdfg5\") pod \"49d542d9-ee75-4f69-a103-0abb25d8c891\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.281038 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"49d542d9-ee75-4f69-a103-0abb25d8c891\" (UID: \"49d542d9-ee75-4f69-a103-0abb25d8c891\") " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.279188 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-lock" (OuterVolumeSpecName: "lock") pod "49d542d9-ee75-4f69-a103-0abb25d8c891" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.281626 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-cache" (OuterVolumeSpecName: "cache") pod "49d542d9-ee75-4f69-a103-0abb25d8c891" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: E1004 03:26:03.282115 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c\": container with ID starting with 51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c not found: ID does not exist" containerID="51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.282153 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c"} err="failed to get container status \"51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c\": rpc error: code = NotFound desc = could not find container \"51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c\": container with ID starting with 51fe8aa155cbe276284d0e3b4d4be2b6adec536620859513e8959a7823f3ea9c not found: ID does not exist" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.282187 4770 scope.go:117] "RemoveContainer" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.282472 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "49d542d9-ee75-4f69-a103-0abb25d8c891" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.282533 4770 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-lock\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.282545 4770 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/49d542d9-ee75-4f69-a103-0abb25d8c891-cache\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.282556 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: E1004 03:26:03.283063 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd\": container with ID starting with 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd not found: ID does not exist" containerID="4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.283131 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd"} err="failed to get container status \"4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd\": rpc error: code = NotFound desc = could not find container \"4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd\": container with ID starting with 4c7d42f8eadd1b42678f405a3b3129fbd948f5a07ef9addd3dfb324bdb80bfbd not found: ID does not exist" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.283158 4770 scope.go:117] "RemoveContainer" containerID="b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4" Oct 04 03:26:03 crc kubenswrapper[4770]: E1004 03:26:03.283536 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4\": container with ID starting with b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4 not found: ID does not exist" containerID="b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.283567 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4"} err="failed to get container status \"b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4\": rpc error: code = NotFound desc = could not find container \"b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4\": container with ID starting with b7f6220cb52a772aec1461360c0b9f2f6137509da4ed9a1d5c61eb7be53847a4 not found: ID does not exist" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.284240 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "swift") pod "49d542d9-ee75-4f69-a103-0abb25d8c891" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.285262 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-kube-api-access-fdfg5" (OuterVolumeSpecName: "kube-api-access-fdfg5") pod "49d542d9-ee75-4f69-a103-0abb25d8c891" (UID: "49d542d9-ee75-4f69-a103-0abb25d8c891"). InnerVolumeSpecName "kube-api-access-fdfg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.384513 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdfg5\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-kube-api-access-fdfg5\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.384879 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.384893 4770 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/49d542d9-ee75-4f69-a103-0abb25d8c891-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.407693 4770 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.428436 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.432837 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.486625 4770 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.690721 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" path="/var/lib/kubelet/pods/1908f1eb-8656-43ee-8532-8b9e4ca8e1c2/volumes" Oct 04 03:26:03 crc kubenswrapper[4770]: I1004 03:26:03.691912 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" path="/var/lib/kubelet/pods/c5f2e918-133c-440e-ab52-2a389ddc9ede/volumes" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.153287 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"49d542d9-ee75-4f69-a103-0abb25d8c891","Type":"ContainerDied","Data":"8e860b016b1756b7cff523f827ad0837c9c9bd3312fbaa71d299a216caa7dc26"} Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.153366 4770 scope.go:117] "RemoveContainer" containerID="345b8c0092ebcc04825e922384f0c8ee2e508d97ba89f66363c61e92ce3d2670" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.153496 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.190889 4770 scope.go:117] "RemoveContainer" containerID="2aea2997649b938d2836a8f46cd25d7caab6d10e8f2aeeac5509198aa88b89ca" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.194266 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.205643 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.227911 4770 scope.go:117] "RemoveContainer" containerID="6385e85337f2e8e8a6703c8fe588bbfb6105b68f63512703a1ed8a03cd155903" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.257355 4770 scope.go:117] "RemoveContainer" containerID="72b2d91a64a79c90616f09e24250e6b65a49ae69602fb773330ae5f11fcb04d4" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.287261 4770 scope.go:117] "RemoveContainer" containerID="4da7895e64c6f0a051ead34f62c7e31342e26f449bd134c7e3529d04b73c2f3d" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.318365 4770 scope.go:117] "RemoveContainer" containerID="6c9a46ab54e3016848961bed1d5d1cc0eeddcff8edf1e007687b96267338090d" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.351501 4770 scope.go:117] "RemoveContainer" containerID="336ce838d7770e81da4c7557a31fa795b90be60a48b3e9b579e1d82a0a17207a" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.379498 4770 scope.go:117] "RemoveContainer" containerID="0bbc28ed4f22995dd014ebc4247f63d2b41fd2841242ce9db1f8c81ed20dd754" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.412813 4770 scope.go:117] "RemoveContainer" containerID="62177bc6b3f5210e61d4fa109ad0edab21794e6c92a8e65133e7ae6f064ed8f9" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.436159 4770 scope.go:117] "RemoveContainer" containerID="7470ca0d05ae46b498189734f3f7a811e3f808a06e71678312140fd5f7c389ea" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.460953 4770 scope.go:117] "RemoveContainer" containerID="013814fda0f11cb6869f76d4caabd04fe6d67c24b613c2843c512e76f378e1d2" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.479877 4770 scope.go:117] "RemoveContainer" containerID="8f1dd12bc897fdadff2fb989c2bead87e41df04b90e728e0cf683bc33de025c4" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.499752 4770 scope.go:117] "RemoveContainer" containerID="290870635c010a1c742ef38d88d7a95bc4bda2febe2172b79b9d8b8044db9b56" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.517337 4770 scope.go:117] "RemoveContainer" containerID="804fe60a8b7c232d84bf3cd9ed92ad47513e9d616208774db899e1a670a8fa9d" Oct 04 03:26:04 crc kubenswrapper[4770]: I1004 03:26:04.536328 4770 scope.go:117] "RemoveContainer" containerID="f824e75e4b4bc8b5e89c5df2fd7e73c49c2760133924bc91336e17d29f37ffa6" Oct 04 03:26:05 crc kubenswrapper[4770]: I1004 03:26:05.686985 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" path="/var/lib/kubelet/pods/49d542d9-ee75-4f69-a103-0abb25d8c891/volumes" Oct 04 03:26:06 crc kubenswrapper[4770]: E1004 03:26:06.342770 4770 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 04 03:26:06 crc kubenswrapper[4770]: E1004 03:26:06.342869 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:38.342849684 +0000 UTC m=+1409.634859396 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-worker-config-data" not found Oct 04 03:26:06 crc kubenswrapper[4770]: E1004 03:26:06.342890 4770 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 04 03:26:06 crc kubenswrapper[4770]: E1004 03:26:06.343046 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data podName:8207573d-3535-4814-aff9-15c30c7600c7 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:38.342969617 +0000 UTC m=+1409.634979359 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data") pod "barbican-worker-576495ccf7-bslzp" (UID: "8207573d-3535-4814-aff9-15c30c7600c7") : secret "barbican-config-data" not found Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.685185 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.749235 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w59hs\" (UniqueName: \"kubernetes.io/projected/8207573d-3535-4814-aff9-15c30c7600c7-kube-api-access-w59hs\") pod \"8207573d-3535-4814-aff9-15c30c7600c7\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.749470 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8207573d-3535-4814-aff9-15c30c7600c7-logs\") pod \"8207573d-3535-4814-aff9-15c30c7600c7\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.749552 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data\") pod \"8207573d-3535-4814-aff9-15c30c7600c7\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.749610 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom\") pod \"8207573d-3535-4814-aff9-15c30c7600c7\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.749698 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-combined-ca-bundle\") pod \"8207573d-3535-4814-aff9-15c30c7600c7\" (UID: \"8207573d-3535-4814-aff9-15c30c7600c7\") " Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.751165 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8207573d-3535-4814-aff9-15c30c7600c7-logs" (OuterVolumeSpecName: "logs") pod "8207573d-3535-4814-aff9-15c30c7600c7" (UID: "8207573d-3535-4814-aff9-15c30c7600c7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.757049 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8207573d-3535-4814-aff9-15c30c7600c7-kube-api-access-w59hs" (OuterVolumeSpecName: "kube-api-access-w59hs") pod "8207573d-3535-4814-aff9-15c30c7600c7" (UID: "8207573d-3535-4814-aff9-15c30c7600c7"). InnerVolumeSpecName "kube-api-access-w59hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.757569 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8207573d-3535-4814-aff9-15c30c7600c7" (UID: "8207573d-3535-4814-aff9-15c30c7600c7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.782988 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8207573d-3535-4814-aff9-15c30c7600c7" (UID: "8207573d-3535-4814-aff9-15c30c7600c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.804915 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data" (OuterVolumeSpecName: "config-data") pod "8207573d-3535-4814-aff9-15c30c7600c7" (UID: "8207573d-3535-4814-aff9-15c30c7600c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.852373 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.852442 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w59hs\" (UniqueName: \"kubernetes.io/projected/8207573d-3535-4814-aff9-15c30c7600c7-kube-api-access-w59hs\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.852474 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8207573d-3535-4814-aff9-15c30c7600c7-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.852499 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:06 crc kubenswrapper[4770]: I1004 03:26:06.852524 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8207573d-3535-4814-aff9-15c30c7600c7-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.205563 4770 generic.go:334] "Generic (PLEG): container finished" podID="8207573d-3535-4814-aff9-15c30c7600c7" containerID="897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b" exitCode=137 Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.205624 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-576495ccf7-bslzp" event={"ID":"8207573d-3535-4814-aff9-15c30c7600c7","Type":"ContainerDied","Data":"897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b"} Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.205706 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-576495ccf7-bslzp" event={"ID":"8207573d-3535-4814-aff9-15c30c7600c7","Type":"ContainerDied","Data":"867e45b18f1a5f2f1163a002ede09e816db61b414739946c55e117f8d21a0d37"} Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.205700 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-576495ccf7-bslzp" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.205730 4770 scope.go:117] "RemoveContainer" containerID="897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.246263 4770 scope.go:117] "RemoveContainer" containerID="6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.249126 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-576495ccf7-bslzp"] Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.254829 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-576495ccf7-bslzp"] Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.273586 4770 scope.go:117] "RemoveContainer" containerID="897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b" Oct 04 03:26:07 crc kubenswrapper[4770]: E1004 03:26:07.274102 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b\": container with ID starting with 897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b not found: ID does not exist" containerID="897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.274164 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b"} err="failed to get container status \"897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b\": rpc error: code = NotFound desc = could not find container \"897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b\": container with ID starting with 897b928573c165f07590d230aa620ce589c20fbd4226680178db27da21bc328b not found: ID does not exist" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.274199 4770 scope.go:117] "RemoveContainer" containerID="6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc" Oct 04 03:26:07 crc kubenswrapper[4770]: E1004 03:26:07.274529 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc\": container with ID starting with 6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc not found: ID does not exist" containerID="6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.274552 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc"} err="failed to get container status \"6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc\": rpc error: code = NotFound desc = could not find container \"6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc\": container with ID starting with 6ec3ce55137ac5a5e88e219820ea11b4b628bbc07c891ae04ebe0d6b4fc838dc not found: ID does not exist" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.685719 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8207573d-3535-4814-aff9-15c30c7600c7" path="/var/lib/kubelet/pods/8207573d-3535-4814-aff9-15c30c7600c7/volumes" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.777631 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.869099 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8mms\" (UniqueName: \"kubernetes.io/projected/9b1ac142-741d-48a1-bf5c-982bbf7bace6-kube-api-access-d8mms\") pod \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.869191 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-combined-ca-bundle\") pod \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.869270 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b1ac142-741d-48a1-bf5c-982bbf7bace6-logs\") pod \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.869305 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data\") pod \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.869408 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data-custom\") pod \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\" (UID: \"9b1ac142-741d-48a1-bf5c-982bbf7bace6\") " Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.870743 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b1ac142-741d-48a1-bf5c-982bbf7bace6-logs" (OuterVolumeSpecName: "logs") pod "9b1ac142-741d-48a1-bf5c-982bbf7bace6" (UID: "9b1ac142-741d-48a1-bf5c-982bbf7bace6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.888615 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9b1ac142-741d-48a1-bf5c-982bbf7bace6" (UID: "9b1ac142-741d-48a1-bf5c-982bbf7bace6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.888890 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b1ac142-741d-48a1-bf5c-982bbf7bace6-kube-api-access-d8mms" (OuterVolumeSpecName: "kube-api-access-d8mms") pod "9b1ac142-741d-48a1-bf5c-982bbf7bace6" (UID: "9b1ac142-741d-48a1-bf5c-982bbf7bace6"). InnerVolumeSpecName "kube-api-access-d8mms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.895632 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b1ac142-741d-48a1-bf5c-982bbf7bace6" (UID: "9b1ac142-741d-48a1-bf5c-982bbf7bace6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.915742 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data" (OuterVolumeSpecName: "config-data") pod "9b1ac142-741d-48a1-bf5c-982bbf7bace6" (UID: "9b1ac142-741d-48a1-bf5c-982bbf7bace6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.971814 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.971883 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8mms\" (UniqueName: \"kubernetes.io/projected/9b1ac142-741d-48a1-bf5c-982bbf7bace6-kube-api-access-d8mms\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.971898 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.971936 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b1ac142-741d-48a1-bf5c-982bbf7bace6-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:07 crc kubenswrapper[4770]: I1004 03:26:07.971950 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b1ac142-741d-48a1-bf5c-982bbf7bace6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.225209 4770 generic.go:334] "Generic (PLEG): container finished" podID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerID="a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1" exitCode=137 Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.225398 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.225399 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" event={"ID":"9b1ac142-741d-48a1-bf5c-982bbf7bace6","Type":"ContainerDied","Data":"a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1"} Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.225510 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-867bd5444b-xx8d4" event={"ID":"9b1ac142-741d-48a1-bf5c-982bbf7bace6","Type":"ContainerDied","Data":"d6a49ed7da5a07ce7333bc8b047f047c4f024bac87ba65085c3f9244b5c2353f"} Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.225554 4770 scope.go:117] "RemoveContainer" containerID="a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1" Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.259137 4770 scope.go:117] "RemoveContainer" containerID="da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97" Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.289843 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-867bd5444b-xx8d4"] Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.300990 4770 scope.go:117] "RemoveContainer" containerID="a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1" Oct 04 03:26:08 crc kubenswrapper[4770]: E1004 03:26:08.302058 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1\": container with ID starting with a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1 not found: ID does not exist" containerID="a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1" Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.302148 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1"} err="failed to get container status \"a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1\": rpc error: code = NotFound desc = could not find container \"a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1\": container with ID starting with a61867608ddc07263e7360f76d514a9e75db00d0a0bf913b71eeb76c47b09ae1 not found: ID does not exist" Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.302926 4770 scope.go:117] "RemoveContainer" containerID="da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97" Oct 04 03:26:08 crc kubenswrapper[4770]: E1004 03:26:08.303477 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97\": container with ID starting with da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97 not found: ID does not exist" containerID="da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97" Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.303653 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97"} err="failed to get container status \"da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97\": rpc error: code = NotFound desc = could not find container \"da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97\": container with ID starting with da96168d305a7ac6db850dabbb05cb53f2b73e626fae0848cac62a08a4522d97 not found: ID does not exist" Oct 04 03:26:08 crc kubenswrapper[4770]: I1004 03:26:08.304446 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-867bd5444b-xx8d4"] Oct 04 03:26:09 crc kubenswrapper[4770]: I1004 03:26:09.688776 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" path="/var/lib/kubelet/pods/9b1ac142-741d-48a1-bf5c-982bbf7bace6/volumes" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.758925 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8dk2r"] Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.759982 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41309df4-53d7-42a8-8a5b-4816cc25fdfb" containerName="memcached" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766734 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="41309df4-53d7-42a8-8a5b-4816cc25fdfb" containerName="memcached" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.766808 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8207573d-3535-4814-aff9-15c30c7600c7" containerName="barbican-worker-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766826 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8207573d-3535-4814-aff9-15c30c7600c7" containerName="barbican-worker-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.766847 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766855 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.766868 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8207573d-3535-4814-aff9-15c30c7600c7" containerName="barbican-worker" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766876 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8207573d-3535-4814-aff9-15c30c7600c7" containerName="barbican-worker" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.766890 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" containerName="cinder-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766898 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" containerName="cinder-api" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.766906 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f9865e-be12-4821-a088-956061126f7b" containerName="placement-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766913 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f9865e-be12-4821-a088-956061126f7b" containerName="placement-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.766925 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e826d9bb-183e-401b-bc8b-6cfe0938a7f6" containerName="nova-cell0-conductor-conductor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766933 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e826d9bb-183e-401b-bc8b-6cfe0938a7f6" containerName="nova-cell0-conductor-conductor" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.766947 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="proxy-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766955 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="proxy-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.766967 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerName="glance-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766975 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerName="glance-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.766985 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerName="barbican-keystone-listener-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.766993 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerName="barbican-keystone-listener-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767044 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="openstack-network-exporter" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767056 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="openstack-network-exporter" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767072 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f43e87-e3d9-4508-8bb1-4fe89c64017f" containerName="mysql-bootstrap" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767086 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f43e87-e3d9-4508-8bb1-4fe89c64017f" containerName="mysql-bootstrap" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767103 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerName="glance-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767113 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerName="glance-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767139 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-server" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767150 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-server" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767174 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767187 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767204 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41a2c8f7-f9d1-44bf-8381-ceea9504a909" containerName="nova-scheduler-scheduler" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767215 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="41a2c8f7-f9d1-44bf-8381-ceea9504a909" containerName="nova-scheduler-scheduler" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767228 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="swift-recon-cron" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767241 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="swift-recon-cron" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767257 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-metadata" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767611 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-metadata" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767670 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerName="barbican-worker-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767680 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerName="barbican-worker-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767691 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="ceilometer-notification-agent" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767700 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="ceilometer-notification-agent" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767710 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-replicator" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767718 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-replicator" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767733 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="rsync" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767740 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="rsync" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767751 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerName="barbican-keystone-listener-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767760 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerName="barbican-keystone-listener-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767772 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d5328f9-e5e9-4ad5-96ff-02442117ff0e" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767781 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d5328f9-e5e9-4ad5-96ff-02442117ff0e" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767794 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerName="glance-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767802 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerName="glance-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767815 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.767823 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.767882 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2def82a8-5927-4ebe-ac87-e8ad232797ee" containerName="rabbitmq" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.768845 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2def82a8-5927-4ebe-ac87-e8ad232797ee" containerName="rabbitmq" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.768870 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee078c16-5858-4c24-b936-11d4aa568f65" containerName="keystone-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.768883 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee078c16-5858-4c24-b936-11d4aa568f65" containerName="keystone-api" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.768893 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerName="barbican-worker" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.768902 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerName="barbican-worker" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.768913 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="ceilometer-central-agent" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.768922 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="ceilometer-central-agent" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.768935 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerName="cinder-scheduler" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.768944 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerName="cinder-scheduler" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.768954 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerName="neutron-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.768963 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerName="neutron-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.768977 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af9166c-e2c5-462d-be52-340be8aadfd5" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.768985 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af9166c-e2c5-462d-be52-340be8aadfd5" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.768998 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" containerName="galera" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769028 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" containerName="galera" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769047 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642fe06d-e5b8-461e-a1f7-dbcefc071945" containerName="rabbitmq" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769056 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="642fe06d-e5b8-461e-a1f7-dbcefc071945" containerName="rabbitmq" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769066 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerName="probe" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769074 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerName="probe" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769087 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef58a910-f85e-4e14-95b5-807741097485" containerName="nova-cell1-conductor-conductor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769097 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef58a910-f85e-4e14-95b5-807741097485" containerName="nova-cell1-conductor-conductor" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769114 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-replicator" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769123 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-replicator" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769141 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f43e87-e3d9-4508-8bb1-4fe89c64017f" containerName="galera" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769150 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f43e87-e3d9-4508-8bb1-4fe89c64017f" containerName="galera" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769164 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="ovn-northd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769174 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="ovn-northd" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769187 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerName="neutron-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769195 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerName="neutron-api" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769211 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c00a8476-9758-46f5-864f-9a7b80ca8bd1" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769220 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c00a8476-9758-46f5-864f-9a7b80ca8bd1" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769232 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" containerName="ovn-controller" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769239 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" containerName="ovn-controller" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769250 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769259 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769272 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769279 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769293 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-expirer" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769302 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-expirer" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769313 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server-init" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769320 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server-init" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769333 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-replicator" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769341 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-replicator" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769350 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerName="barbican-keystone-listener" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769357 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerName="barbican-keystone-listener" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769371 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-auditor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769381 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-auditor" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769397 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f9865e-be12-4821-a088-956061126f7b" containerName="placement-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769408 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f9865e-be12-4821-a088-956061126f7b" containerName="placement-api" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769420 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769429 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-api" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769444 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-reaper" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769453 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-reaper" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769460 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-updater" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769468 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-updater" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769480 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerName="barbican-keystone-listener" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769487 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerName="barbican-keystone-listener" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769500 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769508 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769519 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae4b38bb-7bda-45e9-9d87-f79fe72e9008" containerName="kube-state-metrics" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769528 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae4b38bb-7bda-45e9-9d87-f79fe72e9008" containerName="kube-state-metrics" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769538 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2def82a8-5927-4ebe-ac87-e8ad232797ee" containerName="setup-container" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769545 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2def82a8-5927-4ebe-ac87-e8ad232797ee" containerName="setup-container" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769558 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d18c6351-6663-4a3c-84b6-2752f5e66e49" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769566 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d18c6351-6663-4a3c-84b6-2752f5e66e49" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769581 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-server" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769588 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-server" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769600 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769608 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769617 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerName="glance-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769625 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerName="glance-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769635 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-auditor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769642 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-auditor" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769656 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-auditor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769667 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-auditor" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769677 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-updater" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769685 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-updater" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769695 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642fe06d-e5b8-461e-a1f7-dbcefc071945" containerName="setup-container" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769702 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="642fe06d-e5b8-461e-a1f7-dbcefc071945" containerName="setup-container" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769712 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="sg-core" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769720 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="sg-core" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769734 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-server" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769742 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-server" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769751 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" containerName="cinder-api-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769758 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" containerName="cinder-api-log" Oct 04 03:26:57 crc kubenswrapper[4770]: E1004 03:26:57.769766 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" containerName="mysql-bootstrap" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.769774 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" containerName="mysql-bootstrap" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772080 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8207573d-3535-4814-aff9-15c30c7600c7" containerName="barbican-worker-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772105 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772119 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerName="probe" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772134 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-server" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772145 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="ovn-northd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772157 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerName="barbican-keystone-listener" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772166 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="41309df4-53d7-42a8-8a5b-4816cc25fdfb" containerName="memcached" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772282 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-server" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772292 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-updater" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772304 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="sg-core" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772315 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-replicator" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772325 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-auditor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772333 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8207573d-3535-4814-aff9-15c30c7600c7" containerName="barbican-worker" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772342 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d5328f9-e5e9-4ad5-96ff-02442117ff0e" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772353 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerName="neutron-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772366 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-server" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772376 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerName="barbican-keystone-listener" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772388 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovs-vswitchd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772400 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="35f43e87-e3d9-4508-8bb1-4fe89c64017f" containerName="galera" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772408 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-reaper" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772417 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="rsync" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772425 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="proxy-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772440 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae4b38bb-7bda-45e9-9d87-f79fe72e9008" containerName="kube-state-metrics" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772449 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772457 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-metadata" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772468 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2af9166c-e2c5-462d-be52-340be8aadfd5" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772477 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e328a4dc-7a4d-4cb0-82a8-f60df78328d8" containerName="barbican-keystone-listener-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772496 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e826d9bb-183e-401b-bc8b-6cfe0938a7f6" containerName="nova-cell0-conductor-conductor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772523 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dfed2a2-08b2-4df8-ba89-ef3b23cf16f7" containerName="galera" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772563 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1908f1eb-8656-43ee-8532-8b9e4ca8e1c2" containerName="cinder-scheduler" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772582 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef58a910-f85e-4e14-95b5-807741097485" containerName="nova-cell1-conductor-conductor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772594 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-replicator" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772607 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b1ac142-741d-48a1-bf5c-982bbf7bace6" containerName="barbican-keystone-listener-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772619 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" containerName="cinder-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772629 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-auditor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772640 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="container-replicator" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772651 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5f2e918-133c-440e-ab52-2a389ddc9ede" containerName="ovsdb-server" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772661 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d745cd-57f5-4624-9d2c-75353e52fa56" containerName="barbican-api-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772671 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4fdc2b6-62c6-4f1b-aa59-bd44bc7a910c" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772683 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="swift-recon-cron" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772695 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerName="barbican-worker-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772706 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="87293d1f-5e33-4a9a-9978-67487fd14809" containerName="cinder-api-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772716 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="ceilometer-central-agent" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772724 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="145bb94b-ff5d-4b2f-97b1-1b225fc4f164" containerName="neutron-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772735 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f9865e-be12-4821-a088-956061126f7b" containerName="placement-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772744 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c00a8476-9758-46f5-864f-9a7b80ca8bd1" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772755 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerName="glance-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772768 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4f43f8-c9a5-4a80-aab0-2b9226499b27" containerName="ceilometer-notification-agent" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772778 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ab7038-39be-4c09-82ca-00f83aa69e37" containerName="nova-metadata-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772789 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-expirer" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772801 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerName="glance-httpd" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772811 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7916b1d6-76ad-4780-bfc2-2a0a738b8b8d" containerName="ovn-controller" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772826 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2def82a8-5927-4ebe-ac87-e8ad232797ee" containerName="rabbitmq" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772836 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="697bd5d5-22c6-4303-9077-434eaa099f4a" containerName="barbican-worker" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772845 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36939a0-d10b-488e-8bd4-165b1f9ae4e9" containerName="openstack-network-exporter" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772856 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49edb80-7b9b-4b10-afbc-652ec540542e" containerName="glance-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772868 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee078c16-5858-4c24-b936-11d4aa568f65" containerName="keystone-api" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772881 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cb81588-b9f4-4d8a-8baf-33aebaa6e64d" containerName="glance-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772891 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="account-auditor" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772902 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="642fe06d-e5b8-461e-a1f7-dbcefc071945" containerName="rabbitmq" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772914 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d542d9-ee75-4f69-a103-0abb25d8c891" containerName="object-updater" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772928 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="41a2c8f7-f9d1-44bf-8381-ceea9504a909" containerName="nova-scheduler-scheduler" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772938 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d18c6351-6663-4a3c-84b6-2752f5e66e49" containerName="mariadb-account-delete" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772948 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b795408-0144-4d73-8a56-043ce3e38ee6" containerName="nova-api-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.772960 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f9865e-be12-4821-a088-956061126f7b" containerName="placement-log" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.774235 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8dk2r"] Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.774420 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.939877 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-utilities\") pod \"certified-operators-8dk2r\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.939992 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wghrm\" (UniqueName: \"kubernetes.io/projected/8130072d-2ef8-495d-8999-212db1decece-kube-api-access-wghrm\") pod \"certified-operators-8dk2r\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:57 crc kubenswrapper[4770]: I1004 03:26:57.940082 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-catalog-content\") pod \"certified-operators-8dk2r\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.042263 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-utilities\") pod \"certified-operators-8dk2r\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.042395 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wghrm\" (UniqueName: \"kubernetes.io/projected/8130072d-2ef8-495d-8999-212db1decece-kube-api-access-wghrm\") pod \"certified-operators-8dk2r\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.042461 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-catalog-content\") pod \"certified-operators-8dk2r\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.042964 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-utilities\") pod \"certified-operators-8dk2r\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.043392 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-catalog-content\") pod \"certified-operators-8dk2r\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.067894 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wghrm\" (UniqueName: \"kubernetes.io/projected/8130072d-2ef8-495d-8999-212db1decece-kube-api-access-wghrm\") pod \"certified-operators-8dk2r\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.120276 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.623922 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8dk2r"] Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.879982 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dk2r" event={"ID":"8130072d-2ef8-495d-8999-212db1decece","Type":"ContainerStarted","Data":"e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2"} Oct 04 03:26:58 crc kubenswrapper[4770]: I1004 03:26:58.880617 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dk2r" event={"ID":"8130072d-2ef8-495d-8999-212db1decece","Type":"ContainerStarted","Data":"6887ff4fb0cdc6acb5ae66d42debc8bc6a4c3270a39b005a71c8d4275f64aefa"} Oct 04 03:26:59 crc kubenswrapper[4770]: I1004 03:26:59.896612 4770 generic.go:334] "Generic (PLEG): container finished" podID="8130072d-2ef8-495d-8999-212db1decece" containerID="e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2" exitCode=0 Oct 04 03:26:59 crc kubenswrapper[4770]: I1004 03:26:59.896704 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dk2r" event={"ID":"8130072d-2ef8-495d-8999-212db1decece","Type":"ContainerDied","Data":"e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2"} Oct 04 03:27:01 crc kubenswrapper[4770]: I1004 03:27:01.795547 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:27:01 crc kubenswrapper[4770]: I1004 03:27:01.796286 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:27:01 crc kubenswrapper[4770]: I1004 03:27:01.925042 4770 generic.go:334] "Generic (PLEG): container finished" podID="8130072d-2ef8-495d-8999-212db1decece" containerID="b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc" exitCode=0 Oct 04 03:27:01 crc kubenswrapper[4770]: I1004 03:27:01.925092 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dk2r" event={"ID":"8130072d-2ef8-495d-8999-212db1decece","Type":"ContainerDied","Data":"b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc"} Oct 04 03:27:02 crc kubenswrapper[4770]: I1004 03:27:02.935636 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dk2r" event={"ID":"8130072d-2ef8-495d-8999-212db1decece","Type":"ContainerStarted","Data":"6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2"} Oct 04 03:27:02 crc kubenswrapper[4770]: I1004 03:27:02.965748 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8dk2r" podStartSLOduration=3.421444617 podStartE2EDuration="5.965728549s" podCreationTimestamp="2025-10-04 03:26:57 +0000 UTC" firstStartedPulling="2025-10-04 03:26:59.900604808 +0000 UTC m=+1431.192614550" lastFinishedPulling="2025-10-04 03:27:02.44488876 +0000 UTC m=+1433.736898482" observedRunningTime="2025-10-04 03:27:02.956903988 +0000 UTC m=+1434.248913700" watchObservedRunningTime="2025-10-04 03:27:02.965728549 +0000 UTC m=+1434.257738281" Oct 04 03:27:08 crc kubenswrapper[4770]: I1004 03:27:08.121150 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:27:08 crc kubenswrapper[4770]: I1004 03:27:08.121764 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:27:08 crc kubenswrapper[4770]: I1004 03:27:08.177240 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:27:09 crc kubenswrapper[4770]: I1004 03:27:09.051196 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:27:09 crc kubenswrapper[4770]: I1004 03:27:09.108642 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8dk2r"] Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.045668 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8dk2r" podUID="8130072d-2ef8-495d-8999-212db1decece" containerName="registry-server" containerID="cri-o://6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2" gracePeriod=2 Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.474473 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.571033 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-catalog-content\") pod \"8130072d-2ef8-495d-8999-212db1decece\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.571198 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wghrm\" (UniqueName: \"kubernetes.io/projected/8130072d-2ef8-495d-8999-212db1decece-kube-api-access-wghrm\") pod \"8130072d-2ef8-495d-8999-212db1decece\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.571493 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-utilities\") pod \"8130072d-2ef8-495d-8999-212db1decece\" (UID: \"8130072d-2ef8-495d-8999-212db1decece\") " Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.573045 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-utilities" (OuterVolumeSpecName: "utilities") pod "8130072d-2ef8-495d-8999-212db1decece" (UID: "8130072d-2ef8-495d-8999-212db1decece"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.573614 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.577089 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8130072d-2ef8-495d-8999-212db1decece-kube-api-access-wghrm" (OuterVolumeSpecName: "kube-api-access-wghrm") pod "8130072d-2ef8-495d-8999-212db1decece" (UID: "8130072d-2ef8-495d-8999-212db1decece"). InnerVolumeSpecName "kube-api-access-wghrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.627225 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8130072d-2ef8-495d-8999-212db1decece" (UID: "8130072d-2ef8-495d-8999-212db1decece"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.674809 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8130072d-2ef8-495d-8999-212db1decece-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:11 crc kubenswrapper[4770]: I1004 03:27:11.674864 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wghrm\" (UniqueName: \"kubernetes.io/projected/8130072d-2ef8-495d-8999-212db1decece-kube-api-access-wghrm\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.062692 4770 generic.go:334] "Generic (PLEG): container finished" podID="8130072d-2ef8-495d-8999-212db1decece" containerID="6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2" exitCode=0 Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.062736 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dk2r" event={"ID":"8130072d-2ef8-495d-8999-212db1decece","Type":"ContainerDied","Data":"6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2"} Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.062765 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8dk2r" event={"ID":"8130072d-2ef8-495d-8999-212db1decece","Type":"ContainerDied","Data":"6887ff4fb0cdc6acb5ae66d42debc8bc6a4c3270a39b005a71c8d4275f64aefa"} Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.062783 4770 scope.go:117] "RemoveContainer" containerID="6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.062827 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8dk2r" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.102347 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8dk2r"] Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.116200 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8dk2r"] Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.126393 4770 scope.go:117] "RemoveContainer" containerID="b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.155401 4770 scope.go:117] "RemoveContainer" containerID="e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.186952 4770 scope.go:117] "RemoveContainer" containerID="6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2" Oct 04 03:27:12 crc kubenswrapper[4770]: E1004 03:27:12.187726 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2\": container with ID starting with 6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2 not found: ID does not exist" containerID="6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.187803 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2"} err="failed to get container status \"6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2\": rpc error: code = NotFound desc = could not find container \"6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2\": container with ID starting with 6de138a6115e067b30ae618c53f952038b6913d60a3d061e3d5718937c2bc5c2 not found: ID does not exist" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.187857 4770 scope.go:117] "RemoveContainer" containerID="b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc" Oct 04 03:27:12 crc kubenswrapper[4770]: E1004 03:27:12.188380 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc\": container with ID starting with b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc not found: ID does not exist" containerID="b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.188423 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc"} err="failed to get container status \"b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc\": rpc error: code = NotFound desc = could not find container \"b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc\": container with ID starting with b0b15130fc3d5ab65c51482e4b16c45a09d9db4cf115264b478547835d0762dc not found: ID does not exist" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.188454 4770 scope.go:117] "RemoveContainer" containerID="e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2" Oct 04 03:27:12 crc kubenswrapper[4770]: E1004 03:27:12.188961 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2\": container with ID starting with e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2 not found: ID does not exist" containerID="e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2" Oct 04 03:27:12 crc kubenswrapper[4770]: I1004 03:27:12.188993 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2"} err="failed to get container status \"e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2\": rpc error: code = NotFound desc = could not find container \"e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2\": container with ID starting with e85dd23c9d1d46467793f161d468df83164be6b847d30067495becdd72382be2 not found: ID does not exist" Oct 04 03:27:13 crc kubenswrapper[4770]: I1004 03:27:13.691285 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8130072d-2ef8-495d-8999-212db1decece" path="/var/lib/kubelet/pods/8130072d-2ef8-495d-8999-212db1decece/volumes" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.505794 4770 scope.go:117] "RemoveContainer" containerID="970002f335d604db0f240fc84212bcb06e3f17bed2f69add957b8bf4503ff2be" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.544541 4770 scope.go:117] "RemoveContainer" containerID="cfa890ec4ee2fe7f12e80ec7872bbb31140af8442fe957e965fd9230cc298d8d" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.582388 4770 scope.go:117] "RemoveContainer" containerID="35253e9b3bb51aa3385e4b6549e518d5109f6cc1177b4048be361a4ce14ca70b" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.608182 4770 scope.go:117] "RemoveContainer" containerID="48b401920dbda52bc727cb8f3066f846fcf5688b17b9700513de47eeb995f7fb" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.648485 4770 scope.go:117] "RemoveContainer" containerID="2c5f1590663be7eb40c53863ac7c31baca0a8a6b4a313965c166156581e362a2" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.686440 4770 scope.go:117] "RemoveContainer" containerID="3fa4c602bfba3b93aa690acefd5cac4104a43a1685f6fc8bf4e1a15387371281" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.726185 4770 scope.go:117] "RemoveContainer" containerID="1cb8b0950182c8559e48dc79318dc993602cadf6599a8a91e064f0266f789dda" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.756872 4770 scope.go:117] "RemoveContainer" containerID="0218e06d6c52fad79e58c3f1c9b92b86ec1c9c65d56610da75164f0326b481a0" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.787426 4770 scope.go:117] "RemoveContainer" containerID="99abcb9a2d9a291ad93ba6f3aa28f7b9d7d5d9535d25d3f395d236f89a77c628" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.817128 4770 scope.go:117] "RemoveContainer" containerID="f46efdc0729af256ce5b27b4d468a815cdb31c7d04265184e0b0b624114f04fa" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.855127 4770 scope.go:117] "RemoveContainer" containerID="6ca4790905dd98d2b3d8d0601e85120a413fe7048ebdf88f7712f01b13767f88" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.881551 4770 scope.go:117] "RemoveContainer" containerID="78a49b5bce516a915e4c40d15b73c189f051e0f39b9bb11033e9c1a9a3daf2b0" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.910540 4770 scope.go:117] "RemoveContainer" containerID="2c9cd9b93490eaac4e08d158df7a1d401d8e49a8abea19fcf2d103c304bcc073" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.934256 4770 scope.go:117] "RemoveContainer" containerID="3ce2dd1c186025121805ddd2cc61bd1c00a3f693c6b3053acc59af366a0c89f3" Oct 04 03:27:19 crc kubenswrapper[4770]: I1004 03:27:19.964311 4770 scope.go:117] "RemoveContainer" containerID="f386f612a27ebba4fbd5c97ad2193f2f49abf1418ef36ff9c6115064a41b5ab6" Oct 04 03:27:31 crc kubenswrapper[4770]: I1004 03:27:31.796199 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:27:31 crc kubenswrapper[4770]: I1004 03:27:31.796811 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.302267 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-twjhl"] Oct 04 03:27:38 crc kubenswrapper[4770]: E1004 03:27:38.303475 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8130072d-2ef8-495d-8999-212db1decece" containerName="extract-content" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.303498 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8130072d-2ef8-495d-8999-212db1decece" containerName="extract-content" Oct 04 03:27:38 crc kubenswrapper[4770]: E1004 03:27:38.303522 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8130072d-2ef8-495d-8999-212db1decece" containerName="extract-utilities" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.303533 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8130072d-2ef8-495d-8999-212db1decece" containerName="extract-utilities" Oct 04 03:27:38 crc kubenswrapper[4770]: E1004 03:27:38.303551 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8130072d-2ef8-495d-8999-212db1decece" containerName="registry-server" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.303560 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8130072d-2ef8-495d-8999-212db1decece" containerName="registry-server" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.303788 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8130072d-2ef8-495d-8999-212db1decece" containerName="registry-server" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.310444 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.327040 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twjhl"] Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.427331 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-utilities\") pod \"community-operators-twjhl\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.427589 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv6sc\" (UniqueName: \"kubernetes.io/projected/7647e048-e8e4-40fc-86a9-a26c57e691ff-kube-api-access-vv6sc\") pod \"community-operators-twjhl\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.427784 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-catalog-content\") pod \"community-operators-twjhl\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.528748 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv6sc\" (UniqueName: \"kubernetes.io/projected/7647e048-e8e4-40fc-86a9-a26c57e691ff-kube-api-access-vv6sc\") pod \"community-operators-twjhl\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.528844 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-catalog-content\") pod \"community-operators-twjhl\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.528890 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-utilities\") pod \"community-operators-twjhl\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.529399 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-utilities\") pod \"community-operators-twjhl\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.529501 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-catalog-content\") pod \"community-operators-twjhl\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.547962 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv6sc\" (UniqueName: \"kubernetes.io/projected/7647e048-e8e4-40fc-86a9-a26c57e691ff-kube-api-access-vv6sc\") pod \"community-operators-twjhl\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:38 crc kubenswrapper[4770]: I1004 03:27:38.635183 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:39 crc kubenswrapper[4770]: I1004 03:27:39.140193 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twjhl"] Oct 04 03:27:39 crc kubenswrapper[4770]: I1004 03:27:39.386182 4770 generic.go:334] "Generic (PLEG): container finished" podID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerID="7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8" exitCode=0 Oct 04 03:27:39 crc kubenswrapper[4770]: I1004 03:27:39.386248 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twjhl" event={"ID":"7647e048-e8e4-40fc-86a9-a26c57e691ff","Type":"ContainerDied","Data":"7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8"} Oct 04 03:27:39 crc kubenswrapper[4770]: I1004 03:27:39.386297 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twjhl" event={"ID":"7647e048-e8e4-40fc-86a9-a26c57e691ff","Type":"ContainerStarted","Data":"4005eb7a66310286c5bffc3920d40010869714bba7fc68aa144b9fa2a66927e8"} Oct 04 03:27:40 crc kubenswrapper[4770]: I1004 03:27:40.399856 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twjhl" event={"ID":"7647e048-e8e4-40fc-86a9-a26c57e691ff","Type":"ContainerStarted","Data":"7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17"} Oct 04 03:27:41 crc kubenswrapper[4770]: I1004 03:27:41.417140 4770 generic.go:334] "Generic (PLEG): container finished" podID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerID="7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17" exitCode=0 Oct 04 03:27:41 crc kubenswrapper[4770]: I1004 03:27:41.417234 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twjhl" event={"ID":"7647e048-e8e4-40fc-86a9-a26c57e691ff","Type":"ContainerDied","Data":"7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17"} Oct 04 03:27:42 crc kubenswrapper[4770]: I1004 03:27:42.432465 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twjhl" event={"ID":"7647e048-e8e4-40fc-86a9-a26c57e691ff","Type":"ContainerStarted","Data":"1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c"} Oct 04 03:27:42 crc kubenswrapper[4770]: I1004 03:27:42.465317 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-twjhl" podStartSLOduration=2.018781487 podStartE2EDuration="4.465288314s" podCreationTimestamp="2025-10-04 03:27:38 +0000 UTC" firstStartedPulling="2025-10-04 03:27:39.388603681 +0000 UTC m=+1470.680613423" lastFinishedPulling="2025-10-04 03:27:41.835110498 +0000 UTC m=+1473.127120250" observedRunningTime="2025-10-04 03:27:42.458036744 +0000 UTC m=+1473.750046516" watchObservedRunningTime="2025-10-04 03:27:42.465288314 +0000 UTC m=+1473.757298066" Oct 04 03:27:48 crc kubenswrapper[4770]: I1004 03:27:48.635933 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:48 crc kubenswrapper[4770]: I1004 03:27:48.636828 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:48 crc kubenswrapper[4770]: I1004 03:27:48.706188 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:49 crc kubenswrapper[4770]: I1004 03:27:49.573118 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:49 crc kubenswrapper[4770]: I1004 03:27:49.635238 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twjhl"] Oct 04 03:27:51 crc kubenswrapper[4770]: I1004 03:27:51.550403 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-twjhl" podUID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerName="registry-server" containerID="cri-o://1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c" gracePeriod=2 Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.529249 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.566482 4770 generic.go:334] "Generic (PLEG): container finished" podID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerID="1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c" exitCode=0 Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.566545 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twjhl" event={"ID":"7647e048-e8e4-40fc-86a9-a26c57e691ff","Type":"ContainerDied","Data":"1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c"} Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.566598 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twjhl" event={"ID":"7647e048-e8e4-40fc-86a9-a26c57e691ff","Type":"ContainerDied","Data":"4005eb7a66310286c5bffc3920d40010869714bba7fc68aa144b9fa2a66927e8"} Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.566623 4770 scope.go:117] "RemoveContainer" containerID="1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.566730 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twjhl" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.595732 4770 scope.go:117] "RemoveContainer" containerID="7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.625965 4770 scope.go:117] "RemoveContainer" containerID="7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.655935 4770 scope.go:117] "RemoveContainer" containerID="1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c" Oct 04 03:27:52 crc kubenswrapper[4770]: E1004 03:27:52.656354 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c\": container with ID starting with 1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c not found: ID does not exist" containerID="1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.656390 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c"} err="failed to get container status \"1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c\": rpc error: code = NotFound desc = could not find container \"1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c\": container with ID starting with 1b1c78cc0734ee446fd2b1cf7e9382b01fe108edd1d91464524cc80ad658068c not found: ID does not exist" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.656469 4770 scope.go:117] "RemoveContainer" containerID="7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17" Oct 04 03:27:52 crc kubenswrapper[4770]: E1004 03:27:52.657063 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17\": container with ID starting with 7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17 not found: ID does not exist" containerID="7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.657186 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17"} err="failed to get container status \"7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17\": rpc error: code = NotFound desc = could not find container \"7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17\": container with ID starting with 7419ec0cbed1805804d7e1c83babcafc659f32af7e1b207a4d27432141c26a17 not found: ID does not exist" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.657282 4770 scope.go:117] "RemoveContainer" containerID="7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8" Oct 04 03:27:52 crc kubenswrapper[4770]: E1004 03:27:52.657908 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8\": container with ID starting with 7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8 not found: ID does not exist" containerID="7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.657942 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8"} err="failed to get container status \"7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8\": rpc error: code = NotFound desc = could not find container \"7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8\": container with ID starting with 7e3e138697b8972c9b3f76978e9d35a59e893445ac378e52a136361ab84dc8f8 not found: ID does not exist" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.669081 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv6sc\" (UniqueName: \"kubernetes.io/projected/7647e048-e8e4-40fc-86a9-a26c57e691ff-kube-api-access-vv6sc\") pod \"7647e048-e8e4-40fc-86a9-a26c57e691ff\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.669129 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-utilities\") pod \"7647e048-e8e4-40fc-86a9-a26c57e691ff\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.669273 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-catalog-content\") pod \"7647e048-e8e4-40fc-86a9-a26c57e691ff\" (UID: \"7647e048-e8e4-40fc-86a9-a26c57e691ff\") " Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.670300 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-utilities" (OuterVolumeSpecName: "utilities") pod "7647e048-e8e4-40fc-86a9-a26c57e691ff" (UID: "7647e048-e8e4-40fc-86a9-a26c57e691ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.681413 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7647e048-e8e4-40fc-86a9-a26c57e691ff-kube-api-access-vv6sc" (OuterVolumeSpecName: "kube-api-access-vv6sc") pod "7647e048-e8e4-40fc-86a9-a26c57e691ff" (UID: "7647e048-e8e4-40fc-86a9-a26c57e691ff"). InnerVolumeSpecName "kube-api-access-vv6sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.735868 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7647e048-e8e4-40fc-86a9-a26c57e691ff" (UID: "7647e048-e8e4-40fc-86a9-a26c57e691ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.774264 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.774302 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7647e048-e8e4-40fc-86a9-a26c57e691ff-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.774319 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv6sc\" (UniqueName: \"kubernetes.io/projected/7647e048-e8e4-40fc-86a9-a26c57e691ff-kube-api-access-vv6sc\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.919752 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twjhl"] Oct 04 03:27:52 crc kubenswrapper[4770]: I1004 03:27:52.926702 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-twjhl"] Oct 04 03:27:53 crc kubenswrapper[4770]: I1004 03:27:53.689171 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7647e048-e8e4-40fc-86a9-a26c57e691ff" path="/var/lib/kubelet/pods/7647e048-e8e4-40fc-86a9-a26c57e691ff/volumes" Oct 04 03:28:01 crc kubenswrapper[4770]: I1004 03:28:01.795527 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:28:01 crc kubenswrapper[4770]: I1004 03:28:01.798178 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:28:01 crc kubenswrapper[4770]: I1004 03:28:01.798419 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:28:01 crc kubenswrapper[4770]: I1004 03:28:01.799562 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93642b4e9136e576ce1ad898295dea2f212f632a6d746886f2e5f1af8035d45d"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:28:01 crc kubenswrapper[4770]: I1004 03:28:01.799926 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://93642b4e9136e576ce1ad898295dea2f212f632a6d746886f2e5f1af8035d45d" gracePeriod=600 Oct 04 03:28:02 crc kubenswrapper[4770]: I1004 03:28:02.701394 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="93642b4e9136e576ce1ad898295dea2f212f632a6d746886f2e5f1af8035d45d" exitCode=0 Oct 04 03:28:02 crc kubenswrapper[4770]: I1004 03:28:02.701462 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"93642b4e9136e576ce1ad898295dea2f212f632a6d746886f2e5f1af8035d45d"} Oct 04 03:28:02 crc kubenswrapper[4770]: I1004 03:28:02.701991 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a"} Oct 04 03:28:02 crc kubenswrapper[4770]: I1004 03:28:02.702091 4770 scope.go:117] "RemoveContainer" containerID="d6b0b7a6fef39110ff7706ef744d6f617ed372a11387e2977774d987a4194d8f" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.251628 4770 scope.go:117] "RemoveContainer" containerID="1d60905ae32f34af588942deabcb007f319002f38b02f98d03e42cf5c89bca33" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.280927 4770 scope.go:117] "RemoveContainer" containerID="1c08379885b40ec9939052a4eeecf97debed59c25f447022d90963addf4dcf17" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.326126 4770 scope.go:117] "RemoveContainer" containerID="3f304c4c9d4084df4db4c9d1ca2bc9d44285df6bb49699c82b5e8823e8e1ae3f" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.353341 4770 scope.go:117] "RemoveContainer" containerID="2ed2b0c7307cdff4de6c68749e0b18dddacb71011729c291fbd02e14c6b6f756" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.385826 4770 scope.go:117] "RemoveContainer" containerID="ca61c9dff3e5c232108ba39a8b1a8f4f4f5b0eb23bc5838c1ee8c7e06106319e" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.408467 4770 scope.go:117] "RemoveContainer" containerID="d7f95b4d705cde083fc4ef90c09cae5da3e5a4494a1f070421b36175c42dae52" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.433379 4770 scope.go:117] "RemoveContainer" containerID="df651e173b6556cb139982e461ff9c637c9cf1e2459276d87df36652f29f1557" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.474938 4770 scope.go:117] "RemoveContainer" containerID="edfad2a88ea7cdcc0ab69cc5032405bd7fcbabf51fba14a38595d0a638d62106" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.518953 4770 scope.go:117] "RemoveContainer" containerID="5adbac3cda39aabfa24a6562bafcc4eaada11c462c1b8f6a9852131751821625" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.568658 4770 scope.go:117] "RemoveContainer" containerID="c6a84b1e53e4e8d0fb7cc7653661bad2f904e4879e31bc7343a14c27770f22fb" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.612481 4770 scope.go:117] "RemoveContainer" containerID="edf5ef1a49c54d5bdd50958c0f9f2bd3bad10a29cc67fffa8fd87dd4e3b7e29d" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.630025 4770 scope.go:117] "RemoveContainer" containerID="c2e39f05e83ac887adbf3376bdea8800dc2e57fa7efb4f00b67adc9235d689e5" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.669278 4770 scope.go:117] "RemoveContainer" containerID="ba22ef47e0582d9695bafcc9660a1b4adb815a65c5ef9b6fd33f435fbfb67a12" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.701889 4770 scope.go:117] "RemoveContainer" containerID="399ec7d7f30decbc7812a69e724a27e0175509f84cc8cb66a9a7d9ac5b56cd9d" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.748630 4770 scope.go:117] "RemoveContainer" containerID="aff796e9bd542c911e26288c4a0a0e2f7cee9640b0e3b0afe91a5d63dc41800a" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.774713 4770 scope.go:117] "RemoveContainer" containerID="ad1c6f5e9093316bec8bf19c97c384353c2014b68fb11557c892fae059e000b7" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.802967 4770 scope.go:117] "RemoveContainer" containerID="bc0827aed61857d8364429abe1f13dfc75ffc958b4f218ff97327c98f09244f0" Oct 04 03:28:20 crc kubenswrapper[4770]: I1004 03:28:20.828505 4770 scope.go:117] "RemoveContainer" containerID="0259add0c4189d05b42efd063fd05d0d7b7cec99654b19c19844b1ff27eb64c4" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.134406 4770 scope.go:117] "RemoveContainer" containerID="b945952abce048f596643629d0b14f695125f7fc6eddcf6aec9901c2860f9cbb" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.175677 4770 scope.go:117] "RemoveContainer" containerID="7faf1c073ac70ed8106d0926d77195ec92cf251f3043e92547e6382b188686fb" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.206245 4770 scope.go:117] "RemoveContainer" containerID="999a5b5b19028ec70d3c4019c51d43f906d5ba05dfa731441b8d79c2228d7297" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.260118 4770 scope.go:117] "RemoveContainer" containerID="4e30affad8c19685e28c648ecf8ff6c026d5dda9129c4944f788d2e762582b04" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.284901 4770 scope.go:117] "RemoveContainer" containerID="f652f94cc4c1e2df6ea6011817891a4c91174bbbbee764ff5579bda21b7ab8fb" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.312887 4770 scope.go:117] "RemoveContainer" containerID="ea5b04f24ed7e7ce50dead47790400dc0ff130ec57ae8f47fe60ecd17e750aa1" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.334788 4770 scope.go:117] "RemoveContainer" containerID="3944da16b53c7e9a4a5e1b302b1d3a988d3664cb04d82d3a345d8264c57ae1f9" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.364226 4770 scope.go:117] "RemoveContainer" containerID="715a0efbdfc6a5cbb344ca18f4856780c23bb1adeabccdc5c8eda7ca6149c644" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.407569 4770 scope.go:117] "RemoveContainer" containerID="158bcc32260df183555b1a09c48ad264d70fd241b7a7bab2d1ed22f5fd911d0a" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.431094 4770 scope.go:117] "RemoveContainer" containerID="aeb1436e6d02983ff4efd869dcc5b939283d4b9bf4c8232806b9bf33f990a1c6" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.470300 4770 scope.go:117] "RemoveContainer" containerID="1f8470160c79b09e183b8b0818cd1ff1775ade9ca40792f2aa78671ca11f3a5b" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.501724 4770 scope.go:117] "RemoveContainer" containerID="2a04b73716f5487a1c885d411e35e23f8c38353fc1a171192c915f57244bfe0d" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.531073 4770 scope.go:117] "RemoveContainer" containerID="e8c9ef6e4c5a603a6c8991200968c6b52e850056eec488d0768757430caac141" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.566322 4770 scope.go:117] "RemoveContainer" containerID="68b1223ff961d5213ea6bc5690e64d21567c84142a78aaf580e2d2077be0f0d2" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.598193 4770 scope.go:117] "RemoveContainer" containerID="f79f250529ae5445c1fa9f8a7d6af7dcafd0d968c4d30fd4ce6ee7ab033b040a" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.632571 4770 scope.go:117] "RemoveContainer" containerID="2ba074cd36ae8e2080bb5c06e842e5c1124a1cecb16c5a3f5f1ebd17d268f918" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.658505 4770 scope.go:117] "RemoveContainer" containerID="26f529f3c164e5da28d60a276c95d6939aee3b026f95522050e7f3898614759a" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.683857 4770 scope.go:117] "RemoveContainer" containerID="76df273164d25e12be0bd5fba204411c763fba6be089bd2cf529d36352a6edac" Oct 04 03:29:21 crc kubenswrapper[4770]: I1004 03:29:21.705315 4770 scope.go:117] "RemoveContainer" containerID="65e63dbb5b0c9b461f2c4bfb021835089675895e7de625715395587714b49d15" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.162402 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d"] Oct 04 03:30:00 crc kubenswrapper[4770]: E1004 03:30:00.163344 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerName="extract-content" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.163360 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerName="extract-content" Oct 04 03:30:00 crc kubenswrapper[4770]: E1004 03:30:00.163377 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerName="registry-server" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.163384 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerName="registry-server" Oct 04 03:30:00 crc kubenswrapper[4770]: E1004 03:30:00.163402 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerName="extract-utilities" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.163408 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerName="extract-utilities" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.163543 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7647e048-e8e4-40fc-86a9-a26c57e691ff" containerName="registry-server" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.164038 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.170618 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.173196 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.201736 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d"] Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.330477 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46281317-dc20-4bc8-bd56-34683a6903e1-secret-volume\") pod \"collect-profiles-29325810-rtg9d\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.330551 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbszt\" (UniqueName: \"kubernetes.io/projected/46281317-dc20-4bc8-bd56-34683a6903e1-kube-api-access-bbszt\") pod \"collect-profiles-29325810-rtg9d\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.330593 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46281317-dc20-4bc8-bd56-34683a6903e1-config-volume\") pod \"collect-profiles-29325810-rtg9d\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.432310 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46281317-dc20-4bc8-bd56-34683a6903e1-secret-volume\") pod \"collect-profiles-29325810-rtg9d\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.432373 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbszt\" (UniqueName: \"kubernetes.io/projected/46281317-dc20-4bc8-bd56-34683a6903e1-kube-api-access-bbszt\") pod \"collect-profiles-29325810-rtg9d\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.432398 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46281317-dc20-4bc8-bd56-34683a6903e1-config-volume\") pod \"collect-profiles-29325810-rtg9d\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.434362 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46281317-dc20-4bc8-bd56-34683a6903e1-config-volume\") pod \"collect-profiles-29325810-rtg9d\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.454417 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46281317-dc20-4bc8-bd56-34683a6903e1-secret-volume\") pod \"collect-profiles-29325810-rtg9d\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.458059 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbszt\" (UniqueName: \"kubernetes.io/projected/46281317-dc20-4bc8-bd56-34683a6903e1-kube-api-access-bbszt\") pod \"collect-profiles-29325810-rtg9d\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.488435 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.723269 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d"] Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.985027 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" event={"ID":"46281317-dc20-4bc8-bd56-34683a6903e1","Type":"ContainerStarted","Data":"8297f167d0a5a48392a32b35e345b2eb18440e6eb1577e963985eb48aa717c69"} Oct 04 03:30:00 crc kubenswrapper[4770]: I1004 03:30:00.985496 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" event={"ID":"46281317-dc20-4bc8-bd56-34683a6903e1","Type":"ContainerStarted","Data":"5beeb0bb2b50a6fad62accc0c5dc377808277ccd00d4da61fb5b77a0e752b724"} Oct 04 03:30:01 crc kubenswrapper[4770]: I1004 03:30:01.016798 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" podStartSLOduration=1.01677767 podStartE2EDuration="1.01677767s" podCreationTimestamp="2025-10-04 03:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:30:01.002556659 +0000 UTC m=+1612.294566391" watchObservedRunningTime="2025-10-04 03:30:01.01677767 +0000 UTC m=+1612.308787392" Oct 04 03:30:01 crc kubenswrapper[4770]: I1004 03:30:01.995151 4770 generic.go:334] "Generic (PLEG): container finished" podID="46281317-dc20-4bc8-bd56-34683a6903e1" containerID="8297f167d0a5a48392a32b35e345b2eb18440e6eb1577e963985eb48aa717c69" exitCode=0 Oct 04 03:30:01 crc kubenswrapper[4770]: I1004 03:30:01.995210 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" event={"ID":"46281317-dc20-4bc8-bd56-34683a6903e1","Type":"ContainerDied","Data":"8297f167d0a5a48392a32b35e345b2eb18440e6eb1577e963985eb48aa717c69"} Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.330047 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.475978 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbszt\" (UniqueName: \"kubernetes.io/projected/46281317-dc20-4bc8-bd56-34683a6903e1-kube-api-access-bbszt\") pod \"46281317-dc20-4bc8-bd56-34683a6903e1\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.476250 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46281317-dc20-4bc8-bd56-34683a6903e1-config-volume\") pod \"46281317-dc20-4bc8-bd56-34683a6903e1\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.476450 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46281317-dc20-4bc8-bd56-34683a6903e1-secret-volume\") pod \"46281317-dc20-4bc8-bd56-34683a6903e1\" (UID: \"46281317-dc20-4bc8-bd56-34683a6903e1\") " Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.477405 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46281317-dc20-4bc8-bd56-34683a6903e1-config-volume" (OuterVolumeSpecName: "config-volume") pod "46281317-dc20-4bc8-bd56-34683a6903e1" (UID: "46281317-dc20-4bc8-bd56-34683a6903e1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.483419 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46281317-dc20-4bc8-bd56-34683a6903e1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "46281317-dc20-4bc8-bd56-34683a6903e1" (UID: "46281317-dc20-4bc8-bd56-34683a6903e1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.484149 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46281317-dc20-4bc8-bd56-34683a6903e1-kube-api-access-bbszt" (OuterVolumeSpecName: "kube-api-access-bbszt") pod "46281317-dc20-4bc8-bd56-34683a6903e1" (UID: "46281317-dc20-4bc8-bd56-34683a6903e1"). InnerVolumeSpecName "kube-api-access-bbszt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.578510 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbszt\" (UniqueName: \"kubernetes.io/projected/46281317-dc20-4bc8-bd56-34683a6903e1-kube-api-access-bbszt\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.578550 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46281317-dc20-4bc8-bd56-34683a6903e1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:03 crc kubenswrapper[4770]: I1004 03:30:03.578561 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46281317-dc20-4bc8-bd56-34683a6903e1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:04 crc kubenswrapper[4770]: I1004 03:30:04.016618 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" event={"ID":"46281317-dc20-4bc8-bd56-34683a6903e1","Type":"ContainerDied","Data":"5beeb0bb2b50a6fad62accc0c5dc377808277ccd00d4da61fb5b77a0e752b724"} Oct 04 03:30:04 crc kubenswrapper[4770]: I1004 03:30:04.017229 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5beeb0bb2b50a6fad62accc0c5dc377808277ccd00d4da61fb5b77a0e752b724" Oct 04 03:30:04 crc kubenswrapper[4770]: I1004 03:30:04.016688 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d" Oct 04 03:30:21 crc kubenswrapper[4770]: I1004 03:30:21.976057 4770 scope.go:117] "RemoveContainer" containerID="901e11a4d982f87671e376de02bae186efd672322f8829ca222a35929dc8ff60" Oct 04 03:30:22 crc kubenswrapper[4770]: I1004 03:30:22.030826 4770 scope.go:117] "RemoveContainer" containerID="a24fe25e59f6e2eaf70b6a4057291cd447499c578bdba6d5c9b4583baff28184" Oct 04 03:30:22 crc kubenswrapper[4770]: I1004 03:30:22.084515 4770 scope.go:117] "RemoveContainer" containerID="aa4d6131e89f7a1867262efd05891e1a3ea0a823b9344ca76a3ed1cb75b2f092" Oct 04 03:30:22 crc kubenswrapper[4770]: I1004 03:30:22.112311 4770 scope.go:117] "RemoveContainer" containerID="136fdbb6c40c5913eff7c84a5bd828978d5112b321c3e6b34d37ea56010e8b07" Oct 04 03:30:31 crc kubenswrapper[4770]: I1004 03:30:31.796255 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:30:31 crc kubenswrapper[4770]: I1004 03:30:31.797124 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:31:01 crc kubenswrapper[4770]: I1004 03:31:01.795502 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:31:01 crc kubenswrapper[4770]: I1004 03:31:01.796480 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:31:22 crc kubenswrapper[4770]: I1004 03:31:22.206534 4770 scope.go:117] "RemoveContainer" containerID="1e4a44749d34e20093fa0ef2448d2cdc6233bb16e1aa47b08bdfb010277a16ae" Oct 04 03:31:22 crc kubenswrapper[4770]: I1004 03:31:22.238404 4770 scope.go:117] "RemoveContainer" containerID="6bba8b83577d71e684e12c6e57b942c931915975b422c9aa9892702880a260fc" Oct 04 03:31:22 crc kubenswrapper[4770]: I1004 03:31:22.265229 4770 scope.go:117] "RemoveContainer" containerID="30a450deeec6b43046f1e4c12cadb70ffef33664a48ce36b69bf552dadbd53c4" Oct 04 03:31:22 crc kubenswrapper[4770]: I1004 03:31:22.285894 4770 scope.go:117] "RemoveContainer" containerID="fa7c4e0199de7a74a6a51a933f598585e135eac2ee7b99b1d3d49fa44c55ad5a" Oct 04 03:31:22 crc kubenswrapper[4770]: I1004 03:31:22.310062 4770 scope.go:117] "RemoveContainer" containerID="2a8d8f9a02c3dd2f0f0adf8b24176660cc839972d5330bc42bef1b971ce6c49b" Oct 04 03:31:22 crc kubenswrapper[4770]: I1004 03:31:22.347496 4770 scope.go:117] "RemoveContainer" containerID="e953a76ab2110b06b5538727038bb150355effbb23a98d4d39a75531a080b3b0" Oct 04 03:31:22 crc kubenswrapper[4770]: I1004 03:31:22.374808 4770 scope.go:117] "RemoveContainer" containerID="fdfbfac6fc788e30fc0d06f0b561b738e5d141700c8da1feb54964edb74a746c" Oct 04 03:31:31 crc kubenswrapper[4770]: I1004 03:31:31.796898 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:31:31 crc kubenswrapper[4770]: I1004 03:31:31.798311 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:31:31 crc kubenswrapper[4770]: I1004 03:31:31.798477 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:31:31 crc kubenswrapper[4770]: I1004 03:31:31.800061 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:31:31 crc kubenswrapper[4770]: I1004 03:31:31.800203 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" gracePeriod=600 Oct 04 03:31:31 crc kubenswrapper[4770]: E1004 03:31:31.936615 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:31:31 crc kubenswrapper[4770]: I1004 03:31:31.952352 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" exitCode=0 Oct 04 03:31:31 crc kubenswrapper[4770]: I1004 03:31:31.952444 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a"} Oct 04 03:31:31 crc kubenswrapper[4770]: I1004 03:31:31.952508 4770 scope.go:117] "RemoveContainer" containerID="93642b4e9136e576ce1ad898295dea2f212f632a6d746886f2e5f1af8035d45d" Oct 04 03:31:31 crc kubenswrapper[4770]: I1004 03:31:31.954935 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:31:31 crc kubenswrapper[4770]: E1004 03:31:31.955801 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:31:42 crc kubenswrapper[4770]: I1004 03:31:42.673836 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:31:42 crc kubenswrapper[4770]: E1004 03:31:42.675214 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:31:54 crc kubenswrapper[4770]: I1004 03:31:54.673883 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:31:54 crc kubenswrapper[4770]: E1004 03:31:54.674802 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:32:05 crc kubenswrapper[4770]: I1004 03:32:05.674439 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:32:05 crc kubenswrapper[4770]: E1004 03:32:05.675333 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:32:18 crc kubenswrapper[4770]: I1004 03:32:18.673602 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:32:18 crc kubenswrapper[4770]: E1004 03:32:18.674555 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:32:22 crc kubenswrapper[4770]: I1004 03:32:22.528079 4770 scope.go:117] "RemoveContainer" containerID="9138162486e2a0c16b6c114810102c36c1c41126ad6f792e839348193d77a579" Oct 04 03:32:22 crc kubenswrapper[4770]: I1004 03:32:22.557911 4770 scope.go:117] "RemoveContainer" containerID="226e7dfe223b3700639cce5022b00883b758fc927a30ced5786418eeb4b2b69a" Oct 04 03:32:22 crc kubenswrapper[4770]: I1004 03:32:22.607250 4770 scope.go:117] "RemoveContainer" containerID="5bef3ded5c2c95e3979dba7e9d32dc622485969d7322f5de145044df6a55817c" Oct 04 03:32:31 crc kubenswrapper[4770]: I1004 03:32:31.675952 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:32:31 crc kubenswrapper[4770]: E1004 03:32:31.677388 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:32:42 crc kubenswrapper[4770]: I1004 03:32:42.675538 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:32:42 crc kubenswrapper[4770]: E1004 03:32:42.676675 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:32:53 crc kubenswrapper[4770]: I1004 03:32:53.674218 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:32:53 crc kubenswrapper[4770]: E1004 03:32:53.675504 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:33:06 crc kubenswrapper[4770]: I1004 03:33:06.674373 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:33:06 crc kubenswrapper[4770]: E1004 03:33:06.675193 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:33:19 crc kubenswrapper[4770]: I1004 03:33:19.679564 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:33:19 crc kubenswrapper[4770]: E1004 03:33:19.680685 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:33:33 crc kubenswrapper[4770]: I1004 03:33:33.673363 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:33:33 crc kubenswrapper[4770]: E1004 03:33:33.674418 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:33:45 crc kubenswrapper[4770]: I1004 03:33:45.674489 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:33:45 crc kubenswrapper[4770]: E1004 03:33:45.675816 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:33:59 crc kubenswrapper[4770]: I1004 03:33:59.681908 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:33:59 crc kubenswrapper[4770]: E1004 03:33:59.683298 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:34:13 crc kubenswrapper[4770]: I1004 03:34:13.674146 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:34:13 crc kubenswrapper[4770]: E1004 03:34:13.675500 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:34:24 crc kubenswrapper[4770]: I1004 03:34:24.674576 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:34:24 crc kubenswrapper[4770]: E1004 03:34:24.675585 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:34:36 crc kubenswrapper[4770]: I1004 03:34:36.674720 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:34:36 crc kubenswrapper[4770]: E1004 03:34:36.676213 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:34:47 crc kubenswrapper[4770]: I1004 03:34:47.673682 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:34:47 crc kubenswrapper[4770]: E1004 03:34:47.674691 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.132524 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vhpwn"] Oct 04 03:34:53 crc kubenswrapper[4770]: E1004 03:34:53.133526 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46281317-dc20-4bc8-bd56-34683a6903e1" containerName="collect-profiles" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.133549 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="46281317-dc20-4bc8-bd56-34683a6903e1" containerName="collect-profiles" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.133860 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="46281317-dc20-4bc8-bd56-34683a6903e1" containerName="collect-profiles" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.135962 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.141777 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhpwn"] Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.291929 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-utilities\") pod \"redhat-operators-vhpwn\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.292591 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8nww\" (UniqueName: \"kubernetes.io/projected/da561fb0-adff-41bb-b45d-083c485100bb-kube-api-access-j8nww\") pod \"redhat-operators-vhpwn\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.292699 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-catalog-content\") pod \"redhat-operators-vhpwn\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.393985 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8nww\" (UniqueName: \"kubernetes.io/projected/da561fb0-adff-41bb-b45d-083c485100bb-kube-api-access-j8nww\") pod \"redhat-operators-vhpwn\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.394113 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-catalog-content\") pod \"redhat-operators-vhpwn\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.394146 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-utilities\") pod \"redhat-operators-vhpwn\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.394770 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-utilities\") pod \"redhat-operators-vhpwn\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.394778 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-catalog-content\") pod \"redhat-operators-vhpwn\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.421906 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8nww\" (UniqueName: \"kubernetes.io/projected/da561fb0-adff-41bb-b45d-083c485100bb-kube-api-access-j8nww\") pod \"redhat-operators-vhpwn\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.510480 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:34:53 crc kubenswrapper[4770]: I1004 03:34:53.974999 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhpwn"] Oct 04 03:34:54 crc kubenswrapper[4770]: I1004 03:34:54.834631 4770 generic.go:334] "Generic (PLEG): container finished" podID="da561fb0-adff-41bb-b45d-083c485100bb" containerID="05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be" exitCode=0 Oct 04 03:34:54 crc kubenswrapper[4770]: I1004 03:34:54.834724 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhpwn" event={"ID":"da561fb0-adff-41bb-b45d-083c485100bb","Type":"ContainerDied","Data":"05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be"} Oct 04 03:34:54 crc kubenswrapper[4770]: I1004 03:34:54.835294 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhpwn" event={"ID":"da561fb0-adff-41bb-b45d-083c485100bb","Type":"ContainerStarted","Data":"a3a905e7eff2ebb177234f5194295cedbc587204f1cc94af854dba995c6d8213"} Oct 04 03:34:54 crc kubenswrapper[4770]: I1004 03:34:54.837857 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:34:56 crc kubenswrapper[4770]: I1004 03:34:56.857280 4770 generic.go:334] "Generic (PLEG): container finished" podID="da561fb0-adff-41bb-b45d-083c485100bb" containerID="3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556" exitCode=0 Oct 04 03:34:56 crc kubenswrapper[4770]: I1004 03:34:56.857347 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhpwn" event={"ID":"da561fb0-adff-41bb-b45d-083c485100bb","Type":"ContainerDied","Data":"3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556"} Oct 04 03:34:57 crc kubenswrapper[4770]: I1004 03:34:57.871436 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhpwn" event={"ID":"da561fb0-adff-41bb-b45d-083c485100bb","Type":"ContainerStarted","Data":"68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc"} Oct 04 03:34:57 crc kubenswrapper[4770]: I1004 03:34:57.907157 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vhpwn" podStartSLOduration=2.487764867 podStartE2EDuration="4.90712635s" podCreationTimestamp="2025-10-04 03:34:53 +0000 UTC" firstStartedPulling="2025-10-04 03:34:54.837465415 +0000 UTC m=+1906.129475157" lastFinishedPulling="2025-10-04 03:34:57.256826888 +0000 UTC m=+1908.548836640" observedRunningTime="2025-10-04 03:34:57.901023184 +0000 UTC m=+1909.193032896" watchObservedRunningTime="2025-10-04 03:34:57.90712635 +0000 UTC m=+1909.199136122" Oct 04 03:34:59 crc kubenswrapper[4770]: I1004 03:34:59.681360 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:34:59 crc kubenswrapper[4770]: E1004 03:34:59.685708 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:35:03 crc kubenswrapper[4770]: I1004 03:35:03.511522 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:35:03 crc kubenswrapper[4770]: I1004 03:35:03.511924 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:35:03 crc kubenswrapper[4770]: I1004 03:35:03.580201 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:35:03 crc kubenswrapper[4770]: I1004 03:35:03.986218 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:35:04 crc kubenswrapper[4770]: I1004 03:35:04.051602 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhpwn"] Oct 04 03:35:05 crc kubenswrapper[4770]: I1004 03:35:05.952438 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vhpwn" podUID="da561fb0-adff-41bb-b45d-083c485100bb" containerName="registry-server" containerID="cri-o://68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc" gracePeriod=2 Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.413357 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.515447 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-utilities\") pod \"da561fb0-adff-41bb-b45d-083c485100bb\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.515964 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8nww\" (UniqueName: \"kubernetes.io/projected/da561fb0-adff-41bb-b45d-083c485100bb-kube-api-access-j8nww\") pod \"da561fb0-adff-41bb-b45d-083c485100bb\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.516132 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-catalog-content\") pod \"da561fb0-adff-41bb-b45d-083c485100bb\" (UID: \"da561fb0-adff-41bb-b45d-083c485100bb\") " Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.516957 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-utilities" (OuterVolumeSpecName: "utilities") pod "da561fb0-adff-41bb-b45d-083c485100bb" (UID: "da561fb0-adff-41bb-b45d-083c485100bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.522621 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da561fb0-adff-41bb-b45d-083c485100bb-kube-api-access-j8nww" (OuterVolumeSpecName: "kube-api-access-j8nww") pod "da561fb0-adff-41bb-b45d-083c485100bb" (UID: "da561fb0-adff-41bb-b45d-083c485100bb"). InnerVolumeSpecName "kube-api-access-j8nww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.618520 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.618580 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8nww\" (UniqueName: \"kubernetes.io/projected/da561fb0-adff-41bb-b45d-083c485100bb-kube-api-access-j8nww\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.962286 4770 generic.go:334] "Generic (PLEG): container finished" podID="da561fb0-adff-41bb-b45d-083c485100bb" containerID="68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc" exitCode=0 Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.962376 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhpwn" Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.962366 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhpwn" event={"ID":"da561fb0-adff-41bb-b45d-083c485100bb","Type":"ContainerDied","Data":"68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc"} Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.962476 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhpwn" event={"ID":"da561fb0-adff-41bb-b45d-083c485100bb","Type":"ContainerDied","Data":"a3a905e7eff2ebb177234f5194295cedbc587204f1cc94af854dba995c6d8213"} Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.962553 4770 scope.go:117] "RemoveContainer" containerID="68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc" Oct 04 03:35:06 crc kubenswrapper[4770]: I1004 03:35:06.992671 4770 scope.go:117] "RemoveContainer" containerID="3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.027873 4770 scope.go:117] "RemoveContainer" containerID="05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.059800 4770 scope.go:117] "RemoveContainer" containerID="68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc" Oct 04 03:35:07 crc kubenswrapper[4770]: E1004 03:35:07.060484 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc\": container with ID starting with 68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc not found: ID does not exist" containerID="68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.060562 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc"} err="failed to get container status \"68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc\": rpc error: code = NotFound desc = could not find container \"68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc\": container with ID starting with 68d9b24e7823d5ab1033b797d024bbc0b711d1652fe12898513370244a6deedc not found: ID does not exist" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.060611 4770 scope.go:117] "RemoveContainer" containerID="3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556" Oct 04 03:35:07 crc kubenswrapper[4770]: E1004 03:35:07.061219 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556\": container with ID starting with 3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556 not found: ID does not exist" containerID="3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.061290 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556"} err="failed to get container status \"3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556\": rpc error: code = NotFound desc = could not find container \"3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556\": container with ID starting with 3890b1b9433942a12ad89a4995a6b2b6e501c486a63c24aa65b478c17ebb2556 not found: ID does not exist" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.061328 4770 scope.go:117] "RemoveContainer" containerID="05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be" Oct 04 03:35:07 crc kubenswrapper[4770]: E1004 03:35:07.061847 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be\": container with ID starting with 05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be not found: ID does not exist" containerID="05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.061897 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be"} err="failed to get container status \"05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be\": rpc error: code = NotFound desc = could not find container \"05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be\": container with ID starting with 05cdab8298dd7d6137ca7bec000aef2f1465b749f3865f2ff15664698ac158be not found: ID does not exist" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.532652 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da561fb0-adff-41bb-b45d-083c485100bb" (UID: "da561fb0-adff-41bb-b45d-083c485100bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.534791 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da561fb0-adff-41bb-b45d-083c485100bb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.604879 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhpwn"] Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.610870 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vhpwn"] Oct 04 03:35:07 crc kubenswrapper[4770]: I1004 03:35:07.686415 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da561fb0-adff-41bb-b45d-083c485100bb" path="/var/lib/kubelet/pods/da561fb0-adff-41bb-b45d-083c485100bb/volumes" Oct 04 03:35:10 crc kubenswrapper[4770]: I1004 03:35:10.673863 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:35:10 crc kubenswrapper[4770]: E1004 03:35:10.674697 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:35:23 crc kubenswrapper[4770]: I1004 03:35:23.673895 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:35:23 crc kubenswrapper[4770]: E1004 03:35:23.675195 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:35:34 crc kubenswrapper[4770]: I1004 03:35:34.674775 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:35:34 crc kubenswrapper[4770]: E1004 03:35:34.675855 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:35:47 crc kubenswrapper[4770]: I1004 03:35:47.673775 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:35:47 crc kubenswrapper[4770]: E1004 03:35:47.674791 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:36:02 crc kubenswrapper[4770]: I1004 03:36:02.674405 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:36:02 crc kubenswrapper[4770]: E1004 03:36:02.675662 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:36:14 crc kubenswrapper[4770]: I1004 03:36:14.674477 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:36:14 crc kubenswrapper[4770]: E1004 03:36:14.675739 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:36:28 crc kubenswrapper[4770]: I1004 03:36:28.674900 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:36:28 crc kubenswrapper[4770]: E1004 03:36:28.676230 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:36:43 crc kubenswrapper[4770]: I1004 03:36:43.674851 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:36:44 crc kubenswrapper[4770]: I1004 03:36:44.889258 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"17d1369e0cc8b228139560688bc5e1a9e95fc6317dfbccb11a9e654ea699e5d3"} Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.644656 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8hq"] Oct 04 03:37:30 crc kubenswrapper[4770]: E1004 03:37:30.645751 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da561fb0-adff-41bb-b45d-083c485100bb" containerName="registry-server" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.645769 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="da561fb0-adff-41bb-b45d-083c485100bb" containerName="registry-server" Oct 04 03:37:30 crc kubenswrapper[4770]: E1004 03:37:30.645795 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da561fb0-adff-41bb-b45d-083c485100bb" containerName="extract-content" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.645804 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="da561fb0-adff-41bb-b45d-083c485100bb" containerName="extract-content" Oct 04 03:37:30 crc kubenswrapper[4770]: E1004 03:37:30.645823 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da561fb0-adff-41bb-b45d-083c485100bb" containerName="extract-utilities" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.645832 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="da561fb0-adff-41bb-b45d-083c485100bb" containerName="extract-utilities" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.646046 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="da561fb0-adff-41bb-b45d-083c485100bb" containerName="registry-server" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.647431 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.663935 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8hq"] Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.757365 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-utilities\") pod \"redhat-marketplace-lj8hq\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.757469 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntlm5\" (UniqueName: \"kubernetes.io/projected/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-kube-api-access-ntlm5\") pod \"redhat-marketplace-lj8hq\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.757765 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-catalog-content\") pod \"redhat-marketplace-lj8hq\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.861392 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntlm5\" (UniqueName: \"kubernetes.io/projected/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-kube-api-access-ntlm5\") pod \"redhat-marketplace-lj8hq\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.861480 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-catalog-content\") pod \"redhat-marketplace-lj8hq\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.861566 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-utilities\") pod \"redhat-marketplace-lj8hq\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.862275 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-utilities\") pod \"redhat-marketplace-lj8hq\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.862903 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-catalog-content\") pod \"redhat-marketplace-lj8hq\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.894384 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntlm5\" (UniqueName: \"kubernetes.io/projected/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-kube-api-access-ntlm5\") pod \"redhat-marketplace-lj8hq\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:30 crc kubenswrapper[4770]: I1004 03:37:30.968536 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:31 crc kubenswrapper[4770]: I1004 03:37:31.433696 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8hq"] Oct 04 03:37:31 crc kubenswrapper[4770]: W1004 03:37:31.446418 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod418af9e7_0f2c_4a3f_a2c5_5b7f7bd5584a.slice/crio-276650cb813d99ef1155396ce107cbcfcc15d71c6a52bd0de2bf1f98cadcb607 WatchSource:0}: Error finding container 276650cb813d99ef1155396ce107cbcfcc15d71c6a52bd0de2bf1f98cadcb607: Status 404 returned error can't find the container with id 276650cb813d99ef1155396ce107cbcfcc15d71c6a52bd0de2bf1f98cadcb607 Oct 04 03:37:32 crc kubenswrapper[4770]: I1004 03:37:32.371459 4770 generic.go:334] "Generic (PLEG): container finished" podID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerID="0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d" exitCode=0 Oct 04 03:37:32 crc kubenswrapper[4770]: I1004 03:37:32.371869 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8hq" event={"ID":"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a","Type":"ContainerDied","Data":"0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d"} Oct 04 03:37:32 crc kubenswrapper[4770]: I1004 03:37:32.372130 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8hq" event={"ID":"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a","Type":"ContainerStarted","Data":"276650cb813d99ef1155396ce107cbcfcc15d71c6a52bd0de2bf1f98cadcb607"} Oct 04 03:37:33 crc kubenswrapper[4770]: I1004 03:37:33.384374 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8hq" event={"ID":"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a","Type":"ContainerStarted","Data":"a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30"} Oct 04 03:37:34 crc kubenswrapper[4770]: I1004 03:37:34.397176 4770 generic.go:334] "Generic (PLEG): container finished" podID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerID="a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30" exitCode=0 Oct 04 03:37:34 crc kubenswrapper[4770]: I1004 03:37:34.397236 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8hq" event={"ID":"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a","Type":"ContainerDied","Data":"a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30"} Oct 04 03:37:35 crc kubenswrapper[4770]: I1004 03:37:35.409879 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8hq" event={"ID":"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a","Type":"ContainerStarted","Data":"9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d"} Oct 04 03:37:35 crc kubenswrapper[4770]: I1004 03:37:35.443364 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lj8hq" podStartSLOduration=2.917107667 podStartE2EDuration="5.443329147s" podCreationTimestamp="2025-10-04 03:37:30 +0000 UTC" firstStartedPulling="2025-10-04 03:37:32.378199208 +0000 UTC m=+2063.670208960" lastFinishedPulling="2025-10-04 03:37:34.904420698 +0000 UTC m=+2066.196430440" observedRunningTime="2025-10-04 03:37:35.439805557 +0000 UTC m=+2066.731815309" watchObservedRunningTime="2025-10-04 03:37:35.443329147 +0000 UTC m=+2066.735338899" Oct 04 03:37:40 crc kubenswrapper[4770]: I1004 03:37:40.970184 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:40 crc kubenswrapper[4770]: I1004 03:37:40.971127 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:41 crc kubenswrapper[4770]: I1004 03:37:41.043294 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:41 crc kubenswrapper[4770]: I1004 03:37:41.556654 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:41 crc kubenswrapper[4770]: I1004 03:37:41.640771 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8hq"] Oct 04 03:37:43 crc kubenswrapper[4770]: I1004 03:37:43.501936 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lj8hq" podUID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerName="registry-server" containerID="cri-o://9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d" gracePeriod=2 Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.076628 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.209306 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-utilities\") pod \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.209432 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-catalog-content\") pod \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.209686 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntlm5\" (UniqueName: \"kubernetes.io/projected/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-kube-api-access-ntlm5\") pod \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\" (UID: \"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a\") " Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.211383 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-utilities" (OuterVolumeSpecName: "utilities") pod "418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" (UID: "418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.211890 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.219365 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-kube-api-access-ntlm5" (OuterVolumeSpecName: "kube-api-access-ntlm5") pod "418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" (UID: "418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a"). InnerVolumeSpecName "kube-api-access-ntlm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.238098 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" (UID: "418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.313893 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntlm5\" (UniqueName: \"kubernetes.io/projected/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-kube-api-access-ntlm5\") on node \"crc\" DevicePath \"\"" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.313959 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.514426 4770 generic.go:334] "Generic (PLEG): container finished" podID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerID="9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d" exitCode=0 Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.514495 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8hq" event={"ID":"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a","Type":"ContainerDied","Data":"9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d"} Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.514525 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj8hq" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.514555 4770 scope.go:117] "RemoveContainer" containerID="9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.514536 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj8hq" event={"ID":"418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a","Type":"ContainerDied","Data":"276650cb813d99ef1155396ce107cbcfcc15d71c6a52bd0de2bf1f98cadcb607"} Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.537044 4770 scope.go:117] "RemoveContainer" containerID="a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.576881 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8hq"] Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.582669 4770 scope.go:117] "RemoveContainer" containerID="0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.589272 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj8hq"] Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.619273 4770 scope.go:117] "RemoveContainer" containerID="9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d" Oct 04 03:37:44 crc kubenswrapper[4770]: E1004 03:37:44.619792 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d\": container with ID starting with 9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d not found: ID does not exist" containerID="9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.619841 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d"} err="failed to get container status \"9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d\": rpc error: code = NotFound desc = could not find container \"9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d\": container with ID starting with 9d786dfdb51d6c7d44e7f59f90c2510b2d8ebda6af87f766cd9397cce291038d not found: ID does not exist" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.619874 4770 scope.go:117] "RemoveContainer" containerID="a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30" Oct 04 03:37:44 crc kubenswrapper[4770]: E1004 03:37:44.620182 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30\": container with ID starting with a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30 not found: ID does not exist" containerID="a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.620219 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30"} err="failed to get container status \"a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30\": rpc error: code = NotFound desc = could not find container \"a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30\": container with ID starting with a22581655a0cd69ea35629591525dd5b8c6b139f525d8b94699d91fbfa5c1c30 not found: ID does not exist" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.620245 4770 scope.go:117] "RemoveContainer" containerID="0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d" Oct 04 03:37:44 crc kubenswrapper[4770]: E1004 03:37:44.620510 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d\": container with ID starting with 0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d not found: ID does not exist" containerID="0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d" Oct 04 03:37:44 crc kubenswrapper[4770]: I1004 03:37:44.620546 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d"} err="failed to get container status \"0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d\": rpc error: code = NotFound desc = could not find container \"0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d\": container with ID starting with 0ddc15935fab93c20ada0da18e8378e14bde0e147f734db4fce05983260ec25d not found: ID does not exist" Oct 04 03:37:45 crc kubenswrapper[4770]: I1004 03:37:45.691886 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" path="/var/lib/kubelet/pods/418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a/volumes" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.562173 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zf8ch"] Oct 04 03:38:10 crc kubenswrapper[4770]: E1004 03:38:10.563085 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerName="extract-utilities" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.563103 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerName="extract-utilities" Oct 04 03:38:10 crc kubenswrapper[4770]: E1004 03:38:10.563141 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerName="registry-server" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.563152 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerName="registry-server" Oct 04 03:38:10 crc kubenswrapper[4770]: E1004 03:38:10.563173 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerName="extract-content" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.563181 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerName="extract-content" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.563409 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="418af9e7-0f2c-4a3f-a2c5-5b7f7bd5584a" containerName="registry-server" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.564848 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.575337 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zf8ch"] Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.655863 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-catalog-content\") pod \"community-operators-zf8ch\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.655987 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-utilities\") pod \"community-operators-zf8ch\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.656023 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjktw\" (UniqueName: \"kubernetes.io/projected/d64ac083-9350-4939-b27a-ad1ded2a8744-kube-api-access-fjktw\") pod \"community-operators-zf8ch\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.757747 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjktw\" (UniqueName: \"kubernetes.io/projected/d64ac083-9350-4939-b27a-ad1ded2a8744-kube-api-access-fjktw\") pod \"community-operators-zf8ch\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.758198 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-utilities\") pod \"community-operators-zf8ch\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.758371 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-catalog-content\") pod \"community-operators-zf8ch\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.758747 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-utilities\") pod \"community-operators-zf8ch\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.758768 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-catalog-content\") pod \"community-operators-zf8ch\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.797235 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjktw\" (UniqueName: \"kubernetes.io/projected/d64ac083-9350-4939-b27a-ad1ded2a8744-kube-api-access-fjktw\") pod \"community-operators-zf8ch\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:10 crc kubenswrapper[4770]: I1004 03:38:10.891623 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.387590 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zf8ch"] Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.558236 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bxpl7"] Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.561102 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.573423 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxpl7"] Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.672289 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-utilities\") pod \"certified-operators-bxpl7\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.672364 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfqlb\" (UniqueName: \"kubernetes.io/projected/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-kube-api-access-dfqlb\") pod \"certified-operators-bxpl7\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.672398 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-catalog-content\") pod \"certified-operators-bxpl7\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.774229 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-utilities\") pod \"certified-operators-bxpl7\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.774342 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfqlb\" (UniqueName: \"kubernetes.io/projected/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-kube-api-access-dfqlb\") pod \"certified-operators-bxpl7\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.774385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-catalog-content\") pod \"certified-operators-bxpl7\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.774963 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-catalog-content\") pod \"certified-operators-bxpl7\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.775745 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-utilities\") pod \"certified-operators-bxpl7\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.779783 4770 generic.go:334] "Generic (PLEG): container finished" podID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerID="98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4" exitCode=0 Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.779833 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf8ch" event={"ID":"d64ac083-9350-4939-b27a-ad1ded2a8744","Type":"ContainerDied","Data":"98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4"} Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.779868 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf8ch" event={"ID":"d64ac083-9350-4939-b27a-ad1ded2a8744","Type":"ContainerStarted","Data":"fb6f55c245d772cd50e784b77abed3c0a8e66d4a7c9a729ae5bd2b3f467ea068"} Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.801271 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfqlb\" (UniqueName: \"kubernetes.io/projected/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-kube-api-access-dfqlb\") pod \"certified-operators-bxpl7\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:11 crc kubenswrapper[4770]: I1004 03:38:11.977616 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:12 crc kubenswrapper[4770]: I1004 03:38:12.279160 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxpl7"] Oct 04 03:38:12 crc kubenswrapper[4770]: I1004 03:38:12.792581 4770 generic.go:334] "Generic (PLEG): container finished" podID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerID="3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f" exitCode=0 Oct 04 03:38:12 crc kubenswrapper[4770]: I1004 03:38:12.792700 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf8ch" event={"ID":"d64ac083-9350-4939-b27a-ad1ded2a8744","Type":"ContainerDied","Data":"3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f"} Oct 04 03:38:12 crc kubenswrapper[4770]: I1004 03:38:12.798326 4770 generic.go:334] "Generic (PLEG): container finished" podID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerID="0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c" exitCode=0 Oct 04 03:38:12 crc kubenswrapper[4770]: I1004 03:38:12.798392 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxpl7" event={"ID":"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b","Type":"ContainerDied","Data":"0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c"} Oct 04 03:38:12 crc kubenswrapper[4770]: I1004 03:38:12.798447 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxpl7" event={"ID":"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b","Type":"ContainerStarted","Data":"328cb281a62530dad4ce476580c3171c1c7abca7cbad1c8f32dffb2449d41870"} Oct 04 03:38:13 crc kubenswrapper[4770]: I1004 03:38:13.805942 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxpl7" event={"ID":"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b","Type":"ContainerStarted","Data":"4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f"} Oct 04 03:38:13 crc kubenswrapper[4770]: I1004 03:38:13.807802 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf8ch" event={"ID":"d64ac083-9350-4939-b27a-ad1ded2a8744","Type":"ContainerStarted","Data":"717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8"} Oct 04 03:38:13 crc kubenswrapper[4770]: I1004 03:38:13.845979 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zf8ch" podStartSLOduration=2.375114251 podStartE2EDuration="3.845959089s" podCreationTimestamp="2025-10-04 03:38:10 +0000 UTC" firstStartedPulling="2025-10-04 03:38:11.781825499 +0000 UTC m=+2103.073835211" lastFinishedPulling="2025-10-04 03:38:13.252670337 +0000 UTC m=+2104.544680049" observedRunningTime="2025-10-04 03:38:13.845170839 +0000 UTC m=+2105.137180581" watchObservedRunningTime="2025-10-04 03:38:13.845959089 +0000 UTC m=+2105.137968811" Oct 04 03:38:14 crc kubenswrapper[4770]: I1004 03:38:14.818639 4770 generic.go:334] "Generic (PLEG): container finished" podID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerID="4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f" exitCode=0 Oct 04 03:38:14 crc kubenswrapper[4770]: I1004 03:38:14.818705 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxpl7" event={"ID":"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b","Type":"ContainerDied","Data":"4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f"} Oct 04 03:38:15 crc kubenswrapper[4770]: I1004 03:38:15.837093 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxpl7" event={"ID":"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b","Type":"ContainerStarted","Data":"1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26"} Oct 04 03:38:15 crc kubenswrapper[4770]: I1004 03:38:15.872743 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bxpl7" podStartSLOduration=2.331738053 podStartE2EDuration="4.872726022s" podCreationTimestamp="2025-10-04 03:38:11 +0000 UTC" firstStartedPulling="2025-10-04 03:38:12.800388178 +0000 UTC m=+2104.092397900" lastFinishedPulling="2025-10-04 03:38:15.341376147 +0000 UTC m=+2106.633385869" observedRunningTime="2025-10-04 03:38:15.870159166 +0000 UTC m=+2107.162168878" watchObservedRunningTime="2025-10-04 03:38:15.872726022 +0000 UTC m=+2107.164735734" Oct 04 03:38:20 crc kubenswrapper[4770]: I1004 03:38:20.892080 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:20 crc kubenswrapper[4770]: I1004 03:38:20.892585 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:20 crc kubenswrapper[4770]: I1004 03:38:20.964845 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:21 crc kubenswrapper[4770]: I1004 03:38:21.966144 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:21 crc kubenswrapper[4770]: I1004 03:38:21.977731 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:21 crc kubenswrapper[4770]: I1004 03:38:21.977805 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:22 crc kubenswrapper[4770]: I1004 03:38:22.046466 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:22 crc kubenswrapper[4770]: I1004 03:38:22.966244 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:23 crc kubenswrapper[4770]: I1004 03:38:23.353646 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zf8ch"] Oct 04 03:38:23 crc kubenswrapper[4770]: I1004 03:38:23.912997 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zf8ch" podUID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerName="registry-server" containerID="cri-o://717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8" gracePeriod=2 Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.356265 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxpl7"] Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.464517 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.581581 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-catalog-content\") pod \"d64ac083-9350-4939-b27a-ad1ded2a8744\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.581703 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjktw\" (UniqueName: \"kubernetes.io/projected/d64ac083-9350-4939-b27a-ad1ded2a8744-kube-api-access-fjktw\") pod \"d64ac083-9350-4939-b27a-ad1ded2a8744\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.581904 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-utilities\") pod \"d64ac083-9350-4939-b27a-ad1ded2a8744\" (UID: \"d64ac083-9350-4939-b27a-ad1ded2a8744\") " Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.582881 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-utilities" (OuterVolumeSpecName: "utilities") pod "d64ac083-9350-4939-b27a-ad1ded2a8744" (UID: "d64ac083-9350-4939-b27a-ad1ded2a8744"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.589639 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64ac083-9350-4939-b27a-ad1ded2a8744-kube-api-access-fjktw" (OuterVolumeSpecName: "kube-api-access-fjktw") pod "d64ac083-9350-4939-b27a-ad1ded2a8744" (UID: "d64ac083-9350-4939-b27a-ad1ded2a8744"). InnerVolumeSpecName "kube-api-access-fjktw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.657779 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d64ac083-9350-4939-b27a-ad1ded2a8744" (UID: "d64ac083-9350-4939-b27a-ad1ded2a8744"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.683933 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.684023 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64ac083-9350-4939-b27a-ad1ded2a8744-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.684046 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjktw\" (UniqueName: \"kubernetes.io/projected/d64ac083-9350-4939-b27a-ad1ded2a8744-kube-api-access-fjktw\") on node \"crc\" DevicePath \"\"" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.922276 4770 generic.go:334] "Generic (PLEG): container finished" podID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerID="717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8" exitCode=0 Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.922348 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf8ch" event={"ID":"d64ac083-9350-4939-b27a-ad1ded2a8744","Type":"ContainerDied","Data":"717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8"} Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.922419 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zf8ch" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.922453 4770 scope.go:117] "RemoveContainer" containerID="717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.922436 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zf8ch" event={"ID":"d64ac083-9350-4939-b27a-ad1ded2a8744","Type":"ContainerDied","Data":"fb6f55c245d772cd50e784b77abed3c0a8e66d4a7c9a729ae5bd2b3f467ea068"} Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.922549 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bxpl7" podUID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerName="registry-server" containerID="cri-o://1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26" gracePeriod=2 Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.965472 4770 scope.go:117] "RemoveContainer" containerID="3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f" Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.970333 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zf8ch"] Oct 04 03:38:24 crc kubenswrapper[4770]: I1004 03:38:24.981972 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zf8ch"] Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.001230 4770 scope.go:117] "RemoveContainer" containerID="98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.025697 4770 scope.go:117] "RemoveContainer" containerID="717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8" Oct 04 03:38:25 crc kubenswrapper[4770]: E1004 03:38:25.027152 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8\": container with ID starting with 717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8 not found: ID does not exist" containerID="717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.027247 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8"} err="failed to get container status \"717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8\": rpc error: code = NotFound desc = could not find container \"717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8\": container with ID starting with 717fe5d3c3867b3ff3a75ab7525f0027f06bc32a45105787126a743f70adbeb8 not found: ID does not exist" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.027306 4770 scope.go:117] "RemoveContainer" containerID="3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f" Oct 04 03:38:25 crc kubenswrapper[4770]: E1004 03:38:25.027981 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f\": container with ID starting with 3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f not found: ID does not exist" containerID="3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.028052 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f"} err="failed to get container status \"3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f\": rpc error: code = NotFound desc = could not find container \"3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f\": container with ID starting with 3e1d5d46e36887cad7ddfdaca692c61f7d757384d3ce11e3a4cffc141c3f0c1f not found: ID does not exist" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.028097 4770 scope.go:117] "RemoveContainer" containerID="98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4" Oct 04 03:38:25 crc kubenswrapper[4770]: E1004 03:38:25.029439 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4\": container with ID starting with 98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4 not found: ID does not exist" containerID="98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.029468 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4"} err="failed to get container status \"98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4\": rpc error: code = NotFound desc = could not find container \"98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4\": container with ID starting with 98207172ca91772d1554e75bf1b201788d891fcc9ad93ea444339d002a001ec4 not found: ID does not exist" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.276842 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.395734 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-utilities\") pod \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.395783 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-catalog-content\") pod \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.395828 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfqlb\" (UniqueName: \"kubernetes.io/projected/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-kube-api-access-dfqlb\") pod \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\" (UID: \"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b\") " Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.397259 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-utilities" (OuterVolumeSpecName: "utilities") pod "b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" (UID: "b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.401481 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-kube-api-access-dfqlb" (OuterVolumeSpecName: "kube-api-access-dfqlb") pod "b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" (UID: "b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b"). InnerVolumeSpecName "kube-api-access-dfqlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.441512 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" (UID: "b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.497365 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.497416 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.497431 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfqlb\" (UniqueName: \"kubernetes.io/projected/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b-kube-api-access-dfqlb\") on node \"crc\" DevicePath \"\"" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.689908 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64ac083-9350-4939-b27a-ad1ded2a8744" path="/var/lib/kubelet/pods/d64ac083-9350-4939-b27a-ad1ded2a8744/volumes" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.931851 4770 generic.go:334] "Generic (PLEG): container finished" podID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerID="1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26" exitCode=0 Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.931931 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxpl7" event={"ID":"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b","Type":"ContainerDied","Data":"1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26"} Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.931965 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxpl7" event={"ID":"b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b","Type":"ContainerDied","Data":"328cb281a62530dad4ce476580c3171c1c7abca7cbad1c8f32dffb2449d41870"} Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.931988 4770 scope.go:117] "RemoveContainer" containerID="1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.932247 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxpl7" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.967233 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxpl7"] Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.971343 4770 scope.go:117] "RemoveContainer" containerID="4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f" Oct 04 03:38:25 crc kubenswrapper[4770]: I1004 03:38:25.989772 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bxpl7"] Oct 04 03:38:26 crc kubenswrapper[4770]: I1004 03:38:26.016581 4770 scope.go:117] "RemoveContainer" containerID="0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c" Oct 04 03:38:26 crc kubenswrapper[4770]: I1004 03:38:26.064238 4770 scope.go:117] "RemoveContainer" containerID="1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26" Oct 04 03:38:26 crc kubenswrapper[4770]: E1004 03:38:26.064819 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26\": container with ID starting with 1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26 not found: ID does not exist" containerID="1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26" Oct 04 03:38:26 crc kubenswrapper[4770]: I1004 03:38:26.064888 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26"} err="failed to get container status \"1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26\": rpc error: code = NotFound desc = could not find container \"1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26\": container with ID starting with 1a77bc46fb0d555879c08029db9ece73921d6267c2b6f646ad0a0dd767794e26 not found: ID does not exist" Oct 04 03:38:26 crc kubenswrapper[4770]: I1004 03:38:26.064916 4770 scope.go:117] "RemoveContainer" containerID="4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f" Oct 04 03:38:26 crc kubenswrapper[4770]: E1004 03:38:26.066532 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f\": container with ID starting with 4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f not found: ID does not exist" containerID="4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f" Oct 04 03:38:26 crc kubenswrapper[4770]: I1004 03:38:26.066564 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f"} err="failed to get container status \"4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f\": rpc error: code = NotFound desc = could not find container \"4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f\": container with ID starting with 4fa9700dd7448fc5c98a0cbc7b4ae375c95b582aedbe9f9ea974f6fa0c62583f not found: ID does not exist" Oct 04 03:38:26 crc kubenswrapper[4770]: I1004 03:38:26.066584 4770 scope.go:117] "RemoveContainer" containerID="0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c" Oct 04 03:38:26 crc kubenswrapper[4770]: E1004 03:38:26.067081 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c\": container with ID starting with 0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c not found: ID does not exist" containerID="0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c" Oct 04 03:38:26 crc kubenswrapper[4770]: I1004 03:38:26.067141 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c"} err="failed to get container status \"0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c\": rpc error: code = NotFound desc = could not find container \"0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c\": container with ID starting with 0fc011c72013842f9f639baa622569992a893c401f7fa082fa1f20a8ba6cab0c not found: ID does not exist" Oct 04 03:38:27 crc kubenswrapper[4770]: I1004 03:38:27.687781 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" path="/var/lib/kubelet/pods/b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b/volumes" Oct 04 03:39:01 crc kubenswrapper[4770]: I1004 03:39:01.795951 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:39:01 crc kubenswrapper[4770]: I1004 03:39:01.796954 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:39:31 crc kubenswrapper[4770]: I1004 03:39:31.796070 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:39:31 crc kubenswrapper[4770]: I1004 03:39:31.796733 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:40:01 crc kubenswrapper[4770]: I1004 03:40:01.796043 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:40:01 crc kubenswrapper[4770]: I1004 03:40:01.796879 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:40:01 crc kubenswrapper[4770]: I1004 03:40:01.796946 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:40:01 crc kubenswrapper[4770]: I1004 03:40:01.797719 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17d1369e0cc8b228139560688bc5e1a9e95fc6317dfbccb11a9e654ea699e5d3"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:40:01 crc kubenswrapper[4770]: I1004 03:40:01.797813 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://17d1369e0cc8b228139560688bc5e1a9e95fc6317dfbccb11a9e654ea699e5d3" gracePeriod=600 Oct 04 03:40:02 crc kubenswrapper[4770]: I1004 03:40:02.807916 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="17d1369e0cc8b228139560688bc5e1a9e95fc6317dfbccb11a9e654ea699e5d3" exitCode=0 Oct 04 03:40:02 crc kubenswrapper[4770]: I1004 03:40:02.808031 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"17d1369e0cc8b228139560688bc5e1a9e95fc6317dfbccb11a9e654ea699e5d3"} Oct 04 03:40:02 crc kubenswrapper[4770]: I1004 03:40:02.808512 4770 scope.go:117] "RemoveContainer" containerID="dbd686841fdfd72f487d06c70ff81929b16cb76244a657cfdfb7081865ef267a" Oct 04 03:40:02 crc kubenswrapper[4770]: I1004 03:40:02.809493 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe"} Oct 04 03:42:31 crc kubenswrapper[4770]: I1004 03:42:31.795638 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:42:31 crc kubenswrapper[4770]: I1004 03:42:31.796201 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:43:01 crc kubenswrapper[4770]: I1004 03:43:01.795831 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:43:01 crc kubenswrapper[4770]: I1004 03:43:01.798967 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:43:31 crc kubenswrapper[4770]: I1004 03:43:31.795974 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:43:31 crc kubenswrapper[4770]: I1004 03:43:31.796639 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:43:31 crc kubenswrapper[4770]: I1004 03:43:31.796695 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:43:31 crc kubenswrapper[4770]: I1004 03:43:31.797513 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:43:31 crc kubenswrapper[4770]: I1004 03:43:31.797591 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" gracePeriod=600 Oct 04 03:43:31 crc kubenswrapper[4770]: E1004 03:43:31.920174 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:43:32 crc kubenswrapper[4770]: I1004 03:43:32.806287 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" exitCode=0 Oct 04 03:43:32 crc kubenswrapper[4770]: I1004 03:43:32.806352 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe"} Oct 04 03:43:32 crc kubenswrapper[4770]: I1004 03:43:32.806699 4770 scope.go:117] "RemoveContainer" containerID="17d1369e0cc8b228139560688bc5e1a9e95fc6317dfbccb11a9e654ea699e5d3" Oct 04 03:43:32 crc kubenswrapper[4770]: I1004 03:43:32.807316 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:43:32 crc kubenswrapper[4770]: E1004 03:43:32.807669 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:43:47 crc kubenswrapper[4770]: I1004 03:43:47.674849 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:43:47 crc kubenswrapper[4770]: E1004 03:43:47.675928 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:43:58 crc kubenswrapper[4770]: I1004 03:43:58.674665 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:43:58 crc kubenswrapper[4770]: E1004 03:43:58.675362 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:44:13 crc kubenswrapper[4770]: I1004 03:44:13.674258 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:44:13 crc kubenswrapper[4770]: E1004 03:44:13.675291 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:44:24 crc kubenswrapper[4770]: I1004 03:44:24.673640 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:44:24 crc kubenswrapper[4770]: E1004 03:44:24.674692 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:44:37 crc kubenswrapper[4770]: I1004 03:44:37.674639 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:44:37 crc kubenswrapper[4770]: E1004 03:44:37.676212 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:44:48 crc kubenswrapper[4770]: I1004 03:44:48.673682 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:44:48 crc kubenswrapper[4770]: E1004 03:44:48.676831 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.153701 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2"] Oct 04 03:45:00 crc kubenswrapper[4770]: E1004 03:45:00.154658 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.154678 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4770]: E1004 03:45:00.154701 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.154711 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4770]: E1004 03:45:00.154729 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.154740 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4770]: E1004 03:45:00.154757 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.154767 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4770]: E1004 03:45:00.154810 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.154820 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4770]: E1004 03:45:00.154833 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.154844 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.155062 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e40b45-5752-48e9-b2cf-5ce4a9f15f1b" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.155089 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64ac083-9350-4939-b27a-ad1ded2a8744" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.155838 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.158736 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.169262 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.174169 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2"] Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.289735 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e497921b-5585-45d0-97a1-1330033db38e-secret-volume\") pod \"collect-profiles-29325825-4chq2\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.289937 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e497921b-5585-45d0-97a1-1330033db38e-config-volume\") pod \"collect-profiles-29325825-4chq2\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.290124 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp27l\" (UniqueName: \"kubernetes.io/projected/e497921b-5585-45d0-97a1-1330033db38e-kube-api-access-kp27l\") pod \"collect-profiles-29325825-4chq2\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.391159 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e497921b-5585-45d0-97a1-1330033db38e-config-volume\") pod \"collect-profiles-29325825-4chq2\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.391387 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp27l\" (UniqueName: \"kubernetes.io/projected/e497921b-5585-45d0-97a1-1330033db38e-kube-api-access-kp27l\") pod \"collect-profiles-29325825-4chq2\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.391909 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e497921b-5585-45d0-97a1-1330033db38e-secret-volume\") pod \"collect-profiles-29325825-4chq2\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.392683 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e497921b-5585-45d0-97a1-1330033db38e-config-volume\") pod \"collect-profiles-29325825-4chq2\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.401185 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e497921b-5585-45d0-97a1-1330033db38e-secret-volume\") pod \"collect-profiles-29325825-4chq2\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.415503 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp27l\" (UniqueName: \"kubernetes.io/projected/e497921b-5585-45d0-97a1-1330033db38e-kube-api-access-kp27l\") pod \"collect-profiles-29325825-4chq2\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.487249 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:00 crc kubenswrapper[4770]: I1004 03:45:00.725626 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2"] Oct 04 03:45:01 crc kubenswrapper[4770]: I1004 03:45:01.634432 4770 generic.go:334] "Generic (PLEG): container finished" podID="e497921b-5585-45d0-97a1-1330033db38e" containerID="5757b45d9ce0e40de8c5a1cb226b51a2cc8b505db4064e90846bd02386b7180b" exitCode=0 Oct 04 03:45:01 crc kubenswrapper[4770]: I1004 03:45:01.634553 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" event={"ID":"e497921b-5585-45d0-97a1-1330033db38e","Type":"ContainerDied","Data":"5757b45d9ce0e40de8c5a1cb226b51a2cc8b505db4064e90846bd02386b7180b"} Oct 04 03:45:01 crc kubenswrapper[4770]: I1004 03:45:01.634757 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" event={"ID":"e497921b-5585-45d0-97a1-1330033db38e","Type":"ContainerStarted","Data":"3f5beaa52260232567cb7d13887973d0d95d4a840781dd317552ed25a7e2a86b"} Oct 04 03:45:02 crc kubenswrapper[4770]: I1004 03:45:02.674117 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:45:02 crc kubenswrapper[4770]: E1004 03:45:02.675101 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.008702 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.131756 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e497921b-5585-45d0-97a1-1330033db38e-secret-volume\") pod \"e497921b-5585-45d0-97a1-1330033db38e\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.131810 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e497921b-5585-45d0-97a1-1330033db38e-config-volume\") pod \"e497921b-5585-45d0-97a1-1330033db38e\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.131921 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp27l\" (UniqueName: \"kubernetes.io/projected/e497921b-5585-45d0-97a1-1330033db38e-kube-api-access-kp27l\") pod \"e497921b-5585-45d0-97a1-1330033db38e\" (UID: \"e497921b-5585-45d0-97a1-1330033db38e\") " Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.132536 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e497921b-5585-45d0-97a1-1330033db38e-config-volume" (OuterVolumeSpecName: "config-volume") pod "e497921b-5585-45d0-97a1-1330033db38e" (UID: "e497921b-5585-45d0-97a1-1330033db38e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.138184 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e497921b-5585-45d0-97a1-1330033db38e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e497921b-5585-45d0-97a1-1330033db38e" (UID: "e497921b-5585-45d0-97a1-1330033db38e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.138192 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e497921b-5585-45d0-97a1-1330033db38e-kube-api-access-kp27l" (OuterVolumeSpecName: "kube-api-access-kp27l") pod "e497921b-5585-45d0-97a1-1330033db38e" (UID: "e497921b-5585-45d0-97a1-1330033db38e"). InnerVolumeSpecName "kube-api-access-kp27l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.234286 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e497921b-5585-45d0-97a1-1330033db38e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.234338 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e497921b-5585-45d0-97a1-1330033db38e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.234356 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp27l\" (UniqueName: \"kubernetes.io/projected/e497921b-5585-45d0-97a1-1330033db38e-kube-api-access-kp27l\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.659682 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" event={"ID":"e497921b-5585-45d0-97a1-1330033db38e","Type":"ContainerDied","Data":"3f5beaa52260232567cb7d13887973d0d95d4a840781dd317552ed25a7e2a86b"} Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.659744 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f5beaa52260232567cb7d13887973d0d95d4a840781dd317552ed25a7e2a86b" Oct 04 03:45:03 crc kubenswrapper[4770]: I1004 03:45:03.659824 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2" Oct 04 03:45:04 crc kubenswrapper[4770]: I1004 03:45:04.106302 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph"] Oct 04 03:45:04 crc kubenswrapper[4770]: I1004 03:45:04.116102 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325780-z72ph"] Oct 04 03:45:05 crc kubenswrapper[4770]: I1004 03:45:05.691545 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e111a15d-a0c1-470b-9723-e4bbcc2265c1" path="/var/lib/kubelet/pods/e111a15d-a0c1-470b-9723-e4bbcc2265c1/volumes" Oct 04 03:45:15 crc kubenswrapper[4770]: I1004 03:45:15.674290 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:45:15 crc kubenswrapper[4770]: E1004 03:45:15.675353 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:45:23 crc kubenswrapper[4770]: I1004 03:45:23.001945 4770 scope.go:117] "RemoveContainer" containerID="fb77398cd17e7d7e9e65217d5b7fe4c1948ed116f2315633ca94948443af1fa2" Oct 04 03:45:30 crc kubenswrapper[4770]: I1004 03:45:30.673830 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:45:30 crc kubenswrapper[4770]: E1004 03:45:30.676770 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.426511 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6hkcm"] Oct 04 03:45:32 crc kubenswrapper[4770]: E1004 03:45:32.427156 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e497921b-5585-45d0-97a1-1330033db38e" containerName="collect-profiles" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.427171 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e497921b-5585-45d0-97a1-1330033db38e" containerName="collect-profiles" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.427386 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e497921b-5585-45d0-97a1-1330033db38e" containerName="collect-profiles" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.434746 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.454859 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6hkcm"] Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.495708 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slqj8\" (UniqueName: \"kubernetes.io/projected/57abe207-de96-4395-bbf6-d42a5769e6ad-kube-api-access-slqj8\") pod \"redhat-operators-6hkcm\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.495838 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-catalog-content\") pod \"redhat-operators-6hkcm\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.495881 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-utilities\") pod \"redhat-operators-6hkcm\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.597143 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-catalog-content\") pod \"redhat-operators-6hkcm\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.597226 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-utilities\") pod \"redhat-operators-6hkcm\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.597317 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slqj8\" (UniqueName: \"kubernetes.io/projected/57abe207-de96-4395-bbf6-d42a5769e6ad-kube-api-access-slqj8\") pod \"redhat-operators-6hkcm\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.597870 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-utilities\") pod \"redhat-operators-6hkcm\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.598080 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-catalog-content\") pod \"redhat-operators-6hkcm\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.627761 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slqj8\" (UniqueName: \"kubernetes.io/projected/57abe207-de96-4395-bbf6-d42a5769e6ad-kube-api-access-slqj8\") pod \"redhat-operators-6hkcm\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:32 crc kubenswrapper[4770]: I1004 03:45:32.769645 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:33 crc kubenswrapper[4770]: I1004 03:45:33.241306 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6hkcm"] Oct 04 03:45:33 crc kubenswrapper[4770]: I1004 03:45:33.946678 4770 generic.go:334] "Generic (PLEG): container finished" podID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerID="95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0" exitCode=0 Oct 04 03:45:33 crc kubenswrapper[4770]: I1004 03:45:33.946762 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hkcm" event={"ID":"57abe207-de96-4395-bbf6-d42a5769e6ad","Type":"ContainerDied","Data":"95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0"} Oct 04 03:45:33 crc kubenswrapper[4770]: I1004 03:45:33.947140 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hkcm" event={"ID":"57abe207-de96-4395-bbf6-d42a5769e6ad","Type":"ContainerStarted","Data":"f450ca8df0abd0924a252ccd54de8883ffeda633193d97b774b5e06618e83e22"} Oct 04 03:45:33 crc kubenswrapper[4770]: I1004 03:45:33.949568 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:45:35 crc kubenswrapper[4770]: I1004 03:45:35.974896 4770 generic.go:334] "Generic (PLEG): container finished" podID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerID="c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88" exitCode=0 Oct 04 03:45:35 crc kubenswrapper[4770]: I1004 03:45:35.975036 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hkcm" event={"ID":"57abe207-de96-4395-bbf6-d42a5769e6ad","Type":"ContainerDied","Data":"c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88"} Oct 04 03:45:36 crc kubenswrapper[4770]: I1004 03:45:36.990515 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hkcm" event={"ID":"57abe207-de96-4395-bbf6-d42a5769e6ad","Type":"ContainerStarted","Data":"5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5"} Oct 04 03:45:37 crc kubenswrapper[4770]: I1004 03:45:37.022633 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6hkcm" podStartSLOduration=2.583700038 podStartE2EDuration="5.022610409s" podCreationTimestamp="2025-10-04 03:45:32 +0000 UTC" firstStartedPulling="2025-10-04 03:45:33.949218767 +0000 UTC m=+2545.241228489" lastFinishedPulling="2025-10-04 03:45:36.388129148 +0000 UTC m=+2547.680138860" observedRunningTime="2025-10-04 03:45:37.015059972 +0000 UTC m=+2548.307069704" watchObservedRunningTime="2025-10-04 03:45:37.022610409 +0000 UTC m=+2548.314620131" Oct 04 03:45:41 crc kubenswrapper[4770]: I1004 03:45:41.673993 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:45:41 crc kubenswrapper[4770]: E1004 03:45:41.674712 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:45:42 crc kubenswrapper[4770]: I1004 03:45:42.769867 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:42 crc kubenswrapper[4770]: I1004 03:45:42.770325 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:42 crc kubenswrapper[4770]: I1004 03:45:42.845695 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:43 crc kubenswrapper[4770]: I1004 03:45:43.101658 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:43 crc kubenswrapper[4770]: I1004 03:45:43.164534 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6hkcm"] Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.057720 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6hkcm" podUID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerName="registry-server" containerID="cri-o://5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5" gracePeriod=2 Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.526424 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.606629 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-utilities\") pod \"57abe207-de96-4395-bbf6-d42a5769e6ad\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.606712 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slqj8\" (UniqueName: \"kubernetes.io/projected/57abe207-de96-4395-bbf6-d42a5769e6ad-kube-api-access-slqj8\") pod \"57abe207-de96-4395-bbf6-d42a5769e6ad\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.606795 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-catalog-content\") pod \"57abe207-de96-4395-bbf6-d42a5769e6ad\" (UID: \"57abe207-de96-4395-bbf6-d42a5769e6ad\") " Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.607656 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-utilities" (OuterVolumeSpecName: "utilities") pod "57abe207-de96-4395-bbf6-d42a5769e6ad" (UID: "57abe207-de96-4395-bbf6-d42a5769e6ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.612271 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57abe207-de96-4395-bbf6-d42a5769e6ad-kube-api-access-slqj8" (OuterVolumeSpecName: "kube-api-access-slqj8") pod "57abe207-de96-4395-bbf6-d42a5769e6ad" (UID: "57abe207-de96-4395-bbf6-d42a5769e6ad"). InnerVolumeSpecName "kube-api-access-slqj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.708786 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.708832 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slqj8\" (UniqueName: \"kubernetes.io/projected/57abe207-de96-4395-bbf6-d42a5769e6ad-kube-api-access-slqj8\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.722479 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57abe207-de96-4395-bbf6-d42a5769e6ad" (UID: "57abe207-de96-4395-bbf6-d42a5769e6ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4770]: I1004 03:45:45.810180 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57abe207-de96-4395-bbf6-d42a5769e6ad-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.070423 4770 generic.go:334] "Generic (PLEG): container finished" podID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerID="5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5" exitCode=0 Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.070477 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hkcm" event={"ID":"57abe207-de96-4395-bbf6-d42a5769e6ad","Type":"ContainerDied","Data":"5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5"} Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.070508 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hkcm" event={"ID":"57abe207-de96-4395-bbf6-d42a5769e6ad","Type":"ContainerDied","Data":"f450ca8df0abd0924a252ccd54de8883ffeda633193d97b774b5e06618e83e22"} Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.070530 4770 scope.go:117] "RemoveContainer" containerID="5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.070500 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hkcm" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.105496 4770 scope.go:117] "RemoveContainer" containerID="c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.130168 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6hkcm"] Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.140000 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6hkcm"] Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.145274 4770 scope.go:117] "RemoveContainer" containerID="95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.179862 4770 scope.go:117] "RemoveContainer" containerID="5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5" Oct 04 03:45:46 crc kubenswrapper[4770]: E1004 03:45:46.180630 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5\": container with ID starting with 5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5 not found: ID does not exist" containerID="5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.180688 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5"} err="failed to get container status \"5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5\": rpc error: code = NotFound desc = could not find container \"5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5\": container with ID starting with 5e305bc007b2b7ed13d5590ea1040f8e0e50080a16784153179db1ff299850b5 not found: ID does not exist" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.180725 4770 scope.go:117] "RemoveContainer" containerID="c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88" Oct 04 03:45:46 crc kubenswrapper[4770]: E1004 03:45:46.181385 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88\": container with ID starting with c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88 not found: ID does not exist" containerID="c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.181446 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88"} err="failed to get container status \"c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88\": rpc error: code = NotFound desc = could not find container \"c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88\": container with ID starting with c60fbaf7f0f1938fbc538a08a1704dee76c9dbcc8a606a6cd05c698b32f95b88 not found: ID does not exist" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.181487 4770 scope.go:117] "RemoveContainer" containerID="95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0" Oct 04 03:45:46 crc kubenswrapper[4770]: E1004 03:45:46.182139 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0\": container with ID starting with 95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0 not found: ID does not exist" containerID="95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0" Oct 04 03:45:46 crc kubenswrapper[4770]: I1004 03:45:46.182202 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0"} err="failed to get container status \"95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0\": rpc error: code = NotFound desc = could not find container \"95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0\": container with ID starting with 95ba95ac4f851e8b7a9f542c210b116982171445653de0b9488ebe8963cb10b0 not found: ID does not exist" Oct 04 03:45:47 crc kubenswrapper[4770]: I1004 03:45:47.690529 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57abe207-de96-4395-bbf6-d42a5769e6ad" path="/var/lib/kubelet/pods/57abe207-de96-4395-bbf6-d42a5769e6ad/volumes" Oct 04 03:45:54 crc kubenswrapper[4770]: I1004 03:45:54.674870 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:45:54 crc kubenswrapper[4770]: E1004 03:45:54.675996 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:46:09 crc kubenswrapper[4770]: I1004 03:46:09.685162 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:46:09 crc kubenswrapper[4770]: E1004 03:46:09.686123 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:46:21 crc kubenswrapper[4770]: I1004 03:46:21.674672 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:46:21 crc kubenswrapper[4770]: E1004 03:46:21.675914 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:46:33 crc kubenswrapper[4770]: I1004 03:46:33.674343 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:46:33 crc kubenswrapper[4770]: E1004 03:46:33.678547 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:46:47 crc kubenswrapper[4770]: I1004 03:46:47.674669 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:46:47 crc kubenswrapper[4770]: E1004 03:46:47.675693 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:47:00 crc kubenswrapper[4770]: I1004 03:47:00.674002 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:47:00 crc kubenswrapper[4770]: E1004 03:47:00.674736 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:47:14 crc kubenswrapper[4770]: I1004 03:47:14.674529 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:47:14 crc kubenswrapper[4770]: E1004 03:47:14.675709 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:47:28 crc kubenswrapper[4770]: I1004 03:47:28.674454 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:47:28 crc kubenswrapper[4770]: E1004 03:47:28.675434 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:47:43 crc kubenswrapper[4770]: I1004 03:47:43.673932 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:47:43 crc kubenswrapper[4770]: E1004 03:47:43.675773 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:47:57 crc kubenswrapper[4770]: I1004 03:47:57.674049 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:47:57 crc kubenswrapper[4770]: E1004 03:47:57.675184 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:48:12 crc kubenswrapper[4770]: I1004 03:48:12.673597 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:48:12 crc kubenswrapper[4770]: E1004 03:48:12.674602 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:48:26 crc kubenswrapper[4770]: I1004 03:48:26.674295 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:48:26 crc kubenswrapper[4770]: E1004 03:48:26.675602 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.494518 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5l5sf"] Oct 04 03:48:33 crc kubenswrapper[4770]: E1004 03:48:33.495622 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerName="extract-content" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.495649 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerName="extract-content" Oct 04 03:48:33 crc kubenswrapper[4770]: E1004 03:48:33.495685 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerName="extract-utilities" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.495701 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerName="extract-utilities" Oct 04 03:48:33 crc kubenswrapper[4770]: E1004 03:48:33.495728 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerName="registry-server" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.495743 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerName="registry-server" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.495991 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="57abe207-de96-4395-bbf6-d42a5769e6ad" containerName="registry-server" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.500885 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.523440 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5l5sf"] Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.588842 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbrq4\" (UniqueName: \"kubernetes.io/projected/2cded823-d511-4951-a479-ecf2cdc6707f-kube-api-access-gbrq4\") pod \"community-operators-5l5sf\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.588969 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-catalog-content\") pod \"community-operators-5l5sf\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.589068 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-utilities\") pod \"community-operators-5l5sf\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.689764 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-catalog-content\") pod \"community-operators-5l5sf\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.689831 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-utilities\") pod \"community-operators-5l5sf\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.689867 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbrq4\" (UniqueName: \"kubernetes.io/projected/2cded823-d511-4951-a479-ecf2cdc6707f-kube-api-access-gbrq4\") pod \"community-operators-5l5sf\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.690381 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-utilities\") pod \"community-operators-5l5sf\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.690389 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-catalog-content\") pod \"community-operators-5l5sf\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.708851 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbrq4\" (UniqueName: \"kubernetes.io/projected/2cded823-d511-4951-a479-ecf2cdc6707f-kube-api-access-gbrq4\") pod \"community-operators-5l5sf\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:33 crc kubenswrapper[4770]: I1004 03:48:33.871787 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:34 crc kubenswrapper[4770]: I1004 03:48:34.427180 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5l5sf"] Oct 04 03:48:34 crc kubenswrapper[4770]: W1004 03:48:34.443273 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cded823_d511_4951_a479_ecf2cdc6707f.slice/crio-4a16d8d51718e338adfa308f2ad180747ae96bcf8a74f488727a736cda3d64b1 WatchSource:0}: Error finding container 4a16d8d51718e338adfa308f2ad180747ae96bcf8a74f488727a736cda3d64b1: Status 404 returned error can't find the container with id 4a16d8d51718e338adfa308f2ad180747ae96bcf8a74f488727a736cda3d64b1 Oct 04 03:48:34 crc kubenswrapper[4770]: I1004 03:48:34.673456 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l5sf" event={"ID":"2cded823-d511-4951-a479-ecf2cdc6707f","Type":"ContainerStarted","Data":"4a16d8d51718e338adfa308f2ad180747ae96bcf8a74f488727a736cda3d64b1"} Oct 04 03:48:35 crc kubenswrapper[4770]: I1004 03:48:35.682200 4770 generic.go:334] "Generic (PLEG): container finished" podID="2cded823-d511-4951-a479-ecf2cdc6707f" containerID="27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef" exitCode=0 Oct 04 03:48:35 crc kubenswrapper[4770]: I1004 03:48:35.698755 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l5sf" event={"ID":"2cded823-d511-4951-a479-ecf2cdc6707f","Type":"ContainerDied","Data":"27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef"} Oct 04 03:48:37 crc kubenswrapper[4770]: I1004 03:48:37.712813 4770 generic.go:334] "Generic (PLEG): container finished" podID="2cded823-d511-4951-a479-ecf2cdc6707f" containerID="c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363" exitCode=0 Oct 04 03:48:37 crc kubenswrapper[4770]: I1004 03:48:37.714853 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l5sf" event={"ID":"2cded823-d511-4951-a479-ecf2cdc6707f","Type":"ContainerDied","Data":"c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363"} Oct 04 03:48:38 crc kubenswrapper[4770]: I1004 03:48:38.724575 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l5sf" event={"ID":"2cded823-d511-4951-a479-ecf2cdc6707f","Type":"ContainerStarted","Data":"0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522"} Oct 04 03:48:38 crc kubenswrapper[4770]: I1004 03:48:38.750572 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5l5sf" podStartSLOduration=2.931759739 podStartE2EDuration="5.750546642s" podCreationTimestamp="2025-10-04 03:48:33 +0000 UTC" firstStartedPulling="2025-10-04 03:48:35.683820073 +0000 UTC m=+2726.975829785" lastFinishedPulling="2025-10-04 03:48:38.502606966 +0000 UTC m=+2729.794616688" observedRunningTime="2025-10-04 03:48:38.744676449 +0000 UTC m=+2730.036686171" watchObservedRunningTime="2025-10-04 03:48:38.750546642 +0000 UTC m=+2730.042556394" Oct 04 03:48:40 crc kubenswrapper[4770]: I1004 03:48:40.863289 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cfgbw"] Oct 04 03:48:40 crc kubenswrapper[4770]: I1004 03:48:40.865998 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:40 crc kubenswrapper[4770]: I1004 03:48:40.884567 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cfgbw"] Oct 04 03:48:40 crc kubenswrapper[4770]: I1004 03:48:40.915776 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62wjr\" (UniqueName: \"kubernetes.io/projected/a265d9b5-b308-4939-98fc-07b6b6cc2561-kube-api-access-62wjr\") pod \"certified-operators-cfgbw\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:40 crc kubenswrapper[4770]: I1004 03:48:40.915876 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-catalog-content\") pod \"certified-operators-cfgbw\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:40 crc kubenswrapper[4770]: I1004 03:48:40.915963 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-utilities\") pod \"certified-operators-cfgbw\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:41 crc kubenswrapper[4770]: I1004 03:48:41.017252 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62wjr\" (UniqueName: \"kubernetes.io/projected/a265d9b5-b308-4939-98fc-07b6b6cc2561-kube-api-access-62wjr\") pod \"certified-operators-cfgbw\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:41 crc kubenswrapper[4770]: I1004 03:48:41.017312 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-catalog-content\") pod \"certified-operators-cfgbw\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:41 crc kubenswrapper[4770]: I1004 03:48:41.017354 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-utilities\") pod \"certified-operators-cfgbw\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:41 crc kubenswrapper[4770]: I1004 03:48:41.017942 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-utilities\") pod \"certified-operators-cfgbw\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:41 crc kubenswrapper[4770]: I1004 03:48:41.018608 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-catalog-content\") pod \"certified-operators-cfgbw\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:41 crc kubenswrapper[4770]: I1004 03:48:41.044900 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62wjr\" (UniqueName: \"kubernetes.io/projected/a265d9b5-b308-4939-98fc-07b6b6cc2561-kube-api-access-62wjr\") pod \"certified-operators-cfgbw\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:41 crc kubenswrapper[4770]: I1004 03:48:41.234905 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:41 crc kubenswrapper[4770]: I1004 03:48:41.673468 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:48:42 crc kubenswrapper[4770]: I1004 03:48:42.040558 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cfgbw"] Oct 04 03:48:42 crc kubenswrapper[4770]: W1004 03:48:42.172469 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda265d9b5_b308_4939_98fc_07b6b6cc2561.slice/crio-527711db489fd71bfa5884b27be6c8bedc884f9ae321dcd10bb4a704d1bee55e WatchSource:0}: Error finding container 527711db489fd71bfa5884b27be6c8bedc884f9ae321dcd10bb4a704d1bee55e: Status 404 returned error can't find the container with id 527711db489fd71bfa5884b27be6c8bedc884f9ae321dcd10bb4a704d1bee55e Oct 04 03:48:42 crc kubenswrapper[4770]: I1004 03:48:42.764493 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfgbw" event={"ID":"a265d9b5-b308-4939-98fc-07b6b6cc2561","Type":"ContainerStarted","Data":"527711db489fd71bfa5884b27be6c8bedc884f9ae321dcd10bb4a704d1bee55e"} Oct 04 03:48:43 crc kubenswrapper[4770]: I1004 03:48:43.776317 4770 generic.go:334] "Generic (PLEG): container finished" podID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerID="1c817fadac665d9f78489f1df9ef6e23ec3a81e939d5718a9971ced67e036869" exitCode=0 Oct 04 03:48:43 crc kubenswrapper[4770]: I1004 03:48:43.776393 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfgbw" event={"ID":"a265d9b5-b308-4939-98fc-07b6b6cc2561","Type":"ContainerDied","Data":"1c817fadac665d9f78489f1df9ef6e23ec3a81e939d5718a9971ced67e036869"} Oct 04 03:48:43 crc kubenswrapper[4770]: I1004 03:48:43.780658 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"eb8ca10a41c1b2139ab1071a91dbc916ec93f2edcb474db3d35896efc46be31b"} Oct 04 03:48:43 crc kubenswrapper[4770]: I1004 03:48:43.872027 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:43 crc kubenswrapper[4770]: I1004 03:48:43.872085 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:43 crc kubenswrapper[4770]: I1004 03:48:43.926313 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:44 crc kubenswrapper[4770]: I1004 03:48:44.894815 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:45 crc kubenswrapper[4770]: I1004 03:48:45.841764 4770 generic.go:334] "Generic (PLEG): container finished" podID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerID="a37ced535cc338bde41abf3ff6df85c67a2d6f0d3117a7feb6816873da7ed034" exitCode=0 Oct 04 03:48:45 crc kubenswrapper[4770]: I1004 03:48:45.841822 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfgbw" event={"ID":"a265d9b5-b308-4939-98fc-07b6b6cc2561","Type":"ContainerDied","Data":"a37ced535cc338bde41abf3ff6df85c67a2d6f0d3117a7feb6816873da7ed034"} Oct 04 03:48:46 crc kubenswrapper[4770]: I1004 03:48:46.054946 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5l5sf"] Oct 04 03:48:46 crc kubenswrapper[4770]: I1004 03:48:46.852715 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfgbw" event={"ID":"a265d9b5-b308-4939-98fc-07b6b6cc2561","Type":"ContainerStarted","Data":"3ac67b7a9992d6b867a7dd44b480a46c8a483d5b0d61e98e45aa8d3e94f42036"} Oct 04 03:48:46 crc kubenswrapper[4770]: I1004 03:48:46.853104 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5l5sf" podUID="2cded823-d511-4951-a479-ecf2cdc6707f" containerName="registry-server" containerID="cri-o://0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522" gracePeriod=2 Oct 04 03:48:46 crc kubenswrapper[4770]: I1004 03:48:46.878036 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cfgbw" podStartSLOduration=4.421855178 podStartE2EDuration="6.878022798s" podCreationTimestamp="2025-10-04 03:48:40 +0000 UTC" firstStartedPulling="2025-10-04 03:48:43.779099619 +0000 UTC m=+2735.071109381" lastFinishedPulling="2025-10-04 03:48:46.235267249 +0000 UTC m=+2737.527277001" observedRunningTime="2025-10-04 03:48:46.874110786 +0000 UTC m=+2738.166120508" watchObservedRunningTime="2025-10-04 03:48:46.878022798 +0000 UTC m=+2738.170032500" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.371030 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.471920 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-utilities\") pod \"2cded823-d511-4951-a479-ecf2cdc6707f\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.472182 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-catalog-content\") pod \"2cded823-d511-4951-a479-ecf2cdc6707f\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.472278 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbrq4\" (UniqueName: \"kubernetes.io/projected/2cded823-d511-4951-a479-ecf2cdc6707f-kube-api-access-gbrq4\") pod \"2cded823-d511-4951-a479-ecf2cdc6707f\" (UID: \"2cded823-d511-4951-a479-ecf2cdc6707f\") " Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.472772 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-utilities" (OuterVolumeSpecName: "utilities") pod "2cded823-d511-4951-a479-ecf2cdc6707f" (UID: "2cded823-d511-4951-a479-ecf2cdc6707f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.480259 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cded823-d511-4951-a479-ecf2cdc6707f-kube-api-access-gbrq4" (OuterVolumeSpecName: "kube-api-access-gbrq4") pod "2cded823-d511-4951-a479-ecf2cdc6707f" (UID: "2cded823-d511-4951-a479-ecf2cdc6707f"). InnerVolumeSpecName "kube-api-access-gbrq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.522832 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2cded823-d511-4951-a479-ecf2cdc6707f" (UID: "2cded823-d511-4951-a479-ecf2cdc6707f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.574701 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbrq4\" (UniqueName: \"kubernetes.io/projected/2cded823-d511-4951-a479-ecf2cdc6707f-kube-api-access-gbrq4\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.574764 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.574784 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cded823-d511-4951-a479-ecf2cdc6707f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.863103 4770 generic.go:334] "Generic (PLEG): container finished" podID="2cded823-d511-4951-a479-ecf2cdc6707f" containerID="0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522" exitCode=0 Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.863189 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l5sf" event={"ID":"2cded823-d511-4951-a479-ecf2cdc6707f","Type":"ContainerDied","Data":"0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522"} Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.863212 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5l5sf" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.863239 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l5sf" event={"ID":"2cded823-d511-4951-a479-ecf2cdc6707f","Type":"ContainerDied","Data":"4a16d8d51718e338adfa308f2ad180747ae96bcf8a74f488727a736cda3d64b1"} Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.863298 4770 scope.go:117] "RemoveContainer" containerID="0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.890469 4770 scope.go:117] "RemoveContainer" containerID="c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.897683 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5l5sf"] Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.911742 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5l5sf"] Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.912173 4770 scope.go:117] "RemoveContainer" containerID="27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.935221 4770 scope.go:117] "RemoveContainer" containerID="0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522" Oct 04 03:48:47 crc kubenswrapper[4770]: E1004 03:48:47.935637 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522\": container with ID starting with 0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522 not found: ID does not exist" containerID="0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.935672 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522"} err="failed to get container status \"0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522\": rpc error: code = NotFound desc = could not find container \"0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522\": container with ID starting with 0a2b989e522e7856241c5468913a0d07d9057f47b60f1d4c5df01bfe00683522 not found: ID does not exist" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.935701 4770 scope.go:117] "RemoveContainer" containerID="c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363" Oct 04 03:48:47 crc kubenswrapper[4770]: E1004 03:48:47.936146 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363\": container with ID starting with c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363 not found: ID does not exist" containerID="c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.936209 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363"} err="failed to get container status \"c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363\": rpc error: code = NotFound desc = could not find container \"c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363\": container with ID starting with c89dff8d7bbe71167b90ddce0842f3c907c80c52b78f92a21dece2f4ff65b363 not found: ID does not exist" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.936247 4770 scope.go:117] "RemoveContainer" containerID="27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef" Oct 04 03:48:47 crc kubenswrapper[4770]: E1004 03:48:47.936786 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef\": container with ID starting with 27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef not found: ID does not exist" containerID="27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef" Oct 04 03:48:47 crc kubenswrapper[4770]: I1004 03:48:47.936819 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef"} err="failed to get container status \"27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef\": rpc error: code = NotFound desc = could not find container \"27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef\": container with ID starting with 27fd2827b533d620e1849120a2e8ba4f24459436253f6357cb483ba5611df5ef not found: ID does not exist" Oct 04 03:48:49 crc kubenswrapper[4770]: I1004 03:48:49.694379 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cded823-d511-4951-a479-ecf2cdc6707f" path="/var/lib/kubelet/pods/2cded823-d511-4951-a479-ecf2cdc6707f/volumes" Oct 04 03:48:51 crc kubenswrapper[4770]: I1004 03:48:51.238228 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:51 crc kubenswrapper[4770]: I1004 03:48:51.238285 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:51 crc kubenswrapper[4770]: I1004 03:48:51.310579 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:51 crc kubenswrapper[4770]: I1004 03:48:51.955307 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:52 crc kubenswrapper[4770]: I1004 03:48:52.255190 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cfgbw"] Oct 04 03:48:53 crc kubenswrapper[4770]: I1004 03:48:53.918720 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cfgbw" podUID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerName="registry-server" containerID="cri-o://3ac67b7a9992d6b867a7dd44b480a46c8a483d5b0d61e98e45aa8d3e94f42036" gracePeriod=2 Oct 04 03:48:54 crc kubenswrapper[4770]: I1004 03:48:54.962036 4770 generic.go:334] "Generic (PLEG): container finished" podID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerID="3ac67b7a9992d6b867a7dd44b480a46c8a483d5b0d61e98e45aa8d3e94f42036" exitCode=0 Oct 04 03:48:54 crc kubenswrapper[4770]: I1004 03:48:54.962571 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfgbw" event={"ID":"a265d9b5-b308-4939-98fc-07b6b6cc2561","Type":"ContainerDied","Data":"3ac67b7a9992d6b867a7dd44b480a46c8a483d5b0d61e98e45aa8d3e94f42036"} Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.064478 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.210416 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-catalog-content\") pod \"a265d9b5-b308-4939-98fc-07b6b6cc2561\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.210499 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62wjr\" (UniqueName: \"kubernetes.io/projected/a265d9b5-b308-4939-98fc-07b6b6cc2561-kube-api-access-62wjr\") pod \"a265d9b5-b308-4939-98fc-07b6b6cc2561\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.210600 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-utilities\") pod \"a265d9b5-b308-4939-98fc-07b6b6cc2561\" (UID: \"a265d9b5-b308-4939-98fc-07b6b6cc2561\") " Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.211997 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-utilities" (OuterVolumeSpecName: "utilities") pod "a265d9b5-b308-4939-98fc-07b6b6cc2561" (UID: "a265d9b5-b308-4939-98fc-07b6b6cc2561"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.219902 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a265d9b5-b308-4939-98fc-07b6b6cc2561-kube-api-access-62wjr" (OuterVolumeSpecName: "kube-api-access-62wjr") pod "a265d9b5-b308-4939-98fc-07b6b6cc2561" (UID: "a265d9b5-b308-4939-98fc-07b6b6cc2561"). InnerVolumeSpecName "kube-api-access-62wjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.306445 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a265d9b5-b308-4939-98fc-07b6b6cc2561" (UID: "a265d9b5-b308-4939-98fc-07b6b6cc2561"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.312613 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.312645 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62wjr\" (UniqueName: \"kubernetes.io/projected/a265d9b5-b308-4939-98fc-07b6b6cc2561-kube-api-access-62wjr\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.312659 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a265d9b5-b308-4939-98fc-07b6b6cc2561-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.989973 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cfgbw" event={"ID":"a265d9b5-b308-4939-98fc-07b6b6cc2561","Type":"ContainerDied","Data":"527711db489fd71bfa5884b27be6c8bedc884f9ae321dcd10bb4a704d1bee55e"} Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.990281 4770 scope.go:117] "RemoveContainer" containerID="3ac67b7a9992d6b867a7dd44b480a46c8a483d5b0d61e98e45aa8d3e94f42036" Oct 04 03:48:55 crc kubenswrapper[4770]: I1004 03:48:55.990386 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cfgbw" Oct 04 03:48:56 crc kubenswrapper[4770]: I1004 03:48:56.024193 4770 scope.go:117] "RemoveContainer" containerID="a37ced535cc338bde41abf3ff6df85c67a2d6f0d3117a7feb6816873da7ed034" Oct 04 03:48:56 crc kubenswrapper[4770]: I1004 03:48:56.025043 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cfgbw"] Oct 04 03:48:56 crc kubenswrapper[4770]: I1004 03:48:56.031853 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cfgbw"] Oct 04 03:48:56 crc kubenswrapper[4770]: I1004 03:48:56.047178 4770 scope.go:117] "RemoveContainer" containerID="1c817fadac665d9f78489f1df9ef6e23ec3a81e939d5718a9971ced67e036869" Oct 04 03:48:57 crc kubenswrapper[4770]: I1004 03:48:57.689128 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a265d9b5-b308-4939-98fc-07b6b6cc2561" path="/var/lib/kubelet/pods/a265d9b5-b308-4939-98fc-07b6b6cc2561/volumes" Oct 04 03:51:01 crc kubenswrapper[4770]: I1004 03:51:01.796024 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:51:01 crc kubenswrapper[4770]: I1004 03:51:01.796632 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:51:31 crc kubenswrapper[4770]: I1004 03:51:31.796589 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:51:31 crc kubenswrapper[4770]: I1004 03:51:31.797475 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:52:01 crc kubenswrapper[4770]: I1004 03:52:01.796348 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:52:01 crc kubenswrapper[4770]: I1004 03:52:01.797107 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:52:01 crc kubenswrapper[4770]: I1004 03:52:01.797172 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:52:01 crc kubenswrapper[4770]: I1004 03:52:01.797909 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb8ca10a41c1b2139ab1071a91dbc916ec93f2edcb474db3d35896efc46be31b"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:52:01 crc kubenswrapper[4770]: I1004 03:52:01.798003 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://eb8ca10a41c1b2139ab1071a91dbc916ec93f2edcb474db3d35896efc46be31b" gracePeriod=600 Oct 04 03:52:02 crc kubenswrapper[4770]: I1004 03:52:02.695463 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="eb8ca10a41c1b2139ab1071a91dbc916ec93f2edcb474db3d35896efc46be31b" exitCode=0 Oct 04 03:52:02 crc kubenswrapper[4770]: I1004 03:52:02.695552 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"eb8ca10a41c1b2139ab1071a91dbc916ec93f2edcb474db3d35896efc46be31b"} Oct 04 03:52:02 crc kubenswrapper[4770]: I1004 03:52:02.696276 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d"} Oct 04 03:52:02 crc kubenswrapper[4770]: I1004 03:52:02.696384 4770 scope.go:117] "RemoveContainer" containerID="b77b0328cf22b11df5419c17f17e4c3cf57d2aac414f890d5a8f62a5ea6c9efe" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.303857 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-29dsw"] Oct 04 03:54:24 crc kubenswrapper[4770]: E1004 03:54:24.304954 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cded823-d511-4951-a479-ecf2cdc6707f" containerName="extract-content" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.304975 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cded823-d511-4951-a479-ecf2cdc6707f" containerName="extract-content" Oct 04 03:54:24 crc kubenswrapper[4770]: E1004 03:54:24.305029 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cded823-d511-4951-a479-ecf2cdc6707f" containerName="registry-server" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.305043 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cded823-d511-4951-a479-ecf2cdc6707f" containerName="registry-server" Oct 04 03:54:24 crc kubenswrapper[4770]: E1004 03:54:24.305065 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerName="extract-utilities" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.305078 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerName="extract-utilities" Oct 04 03:54:24 crc kubenswrapper[4770]: E1004 03:54:24.305099 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerName="registry-server" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.305111 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerName="registry-server" Oct 04 03:54:24 crc kubenswrapper[4770]: E1004 03:54:24.305128 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cded823-d511-4951-a479-ecf2cdc6707f" containerName="extract-utilities" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.305141 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cded823-d511-4951-a479-ecf2cdc6707f" containerName="extract-utilities" Oct 04 03:54:24 crc kubenswrapper[4770]: E1004 03:54:24.305170 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerName="extract-content" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.305185 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerName="extract-content" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.305673 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cded823-d511-4951-a479-ecf2cdc6707f" containerName="registry-server" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.305720 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a265d9b5-b308-4939-98fc-07b6b6cc2561" containerName="registry-server" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.308487 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.330836 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29dsw"] Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.438542 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8tr4\" (UniqueName: \"kubernetes.io/projected/71637165-aaba-4a30-abbf-641b66f7ad23-kube-api-access-z8tr4\") pod \"redhat-marketplace-29dsw\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.438612 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-utilities\") pod \"redhat-marketplace-29dsw\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.438749 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-catalog-content\") pod \"redhat-marketplace-29dsw\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.540070 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-utilities\") pod \"redhat-marketplace-29dsw\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.540266 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-catalog-content\") pod \"redhat-marketplace-29dsw\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.540364 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8tr4\" (UniqueName: \"kubernetes.io/projected/71637165-aaba-4a30-abbf-641b66f7ad23-kube-api-access-z8tr4\") pod \"redhat-marketplace-29dsw\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.540910 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-utilities\") pod \"redhat-marketplace-29dsw\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.540950 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-catalog-content\") pod \"redhat-marketplace-29dsw\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.560929 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8tr4\" (UniqueName: \"kubernetes.io/projected/71637165-aaba-4a30-abbf-641b66f7ad23-kube-api-access-z8tr4\") pod \"redhat-marketplace-29dsw\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:24 crc kubenswrapper[4770]: I1004 03:54:24.647250 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:25 crc kubenswrapper[4770]: I1004 03:54:25.135686 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29dsw"] Oct 04 03:54:26 crc kubenswrapper[4770]: I1004 03:54:26.083609 4770 generic.go:334] "Generic (PLEG): container finished" podID="71637165-aaba-4a30-abbf-641b66f7ad23" containerID="b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654" exitCode=0 Oct 04 03:54:26 crc kubenswrapper[4770]: I1004 03:54:26.083698 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29dsw" event={"ID":"71637165-aaba-4a30-abbf-641b66f7ad23","Type":"ContainerDied","Data":"b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654"} Oct 04 03:54:26 crc kubenswrapper[4770]: I1004 03:54:26.084298 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29dsw" event={"ID":"71637165-aaba-4a30-abbf-641b66f7ad23","Type":"ContainerStarted","Data":"830c321390890aa496ac01691784b8dbae6b174e44bf94fb8fb313451367ad23"} Oct 04 03:54:26 crc kubenswrapper[4770]: I1004 03:54:26.086517 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:54:27 crc kubenswrapper[4770]: I1004 03:54:27.092233 4770 generic.go:334] "Generic (PLEG): container finished" podID="71637165-aaba-4a30-abbf-641b66f7ad23" containerID="5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784" exitCode=0 Oct 04 03:54:27 crc kubenswrapper[4770]: I1004 03:54:27.092329 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29dsw" event={"ID":"71637165-aaba-4a30-abbf-641b66f7ad23","Type":"ContainerDied","Data":"5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784"} Oct 04 03:54:28 crc kubenswrapper[4770]: I1004 03:54:28.103482 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29dsw" event={"ID":"71637165-aaba-4a30-abbf-641b66f7ad23","Type":"ContainerStarted","Data":"758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7"} Oct 04 03:54:28 crc kubenswrapper[4770]: I1004 03:54:28.133164 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-29dsw" podStartSLOduration=2.71471155 podStartE2EDuration="4.133132161s" podCreationTimestamp="2025-10-04 03:54:24 +0000 UTC" firstStartedPulling="2025-10-04 03:54:26.086001495 +0000 UTC m=+3077.378011247" lastFinishedPulling="2025-10-04 03:54:27.504422106 +0000 UTC m=+3078.796431858" observedRunningTime="2025-10-04 03:54:28.126637793 +0000 UTC m=+3079.418647555" watchObservedRunningTime="2025-10-04 03:54:28.133132161 +0000 UTC m=+3079.425141913" Oct 04 03:54:31 crc kubenswrapper[4770]: I1004 03:54:31.796240 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:54:31 crc kubenswrapper[4770]: I1004 03:54:31.796505 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:54:34 crc kubenswrapper[4770]: I1004 03:54:34.648289 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:34 crc kubenswrapper[4770]: I1004 03:54:34.649912 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:34 crc kubenswrapper[4770]: I1004 03:54:34.723999 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:35 crc kubenswrapper[4770]: I1004 03:54:35.232401 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:35 crc kubenswrapper[4770]: I1004 03:54:35.303263 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29dsw"] Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.183083 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-29dsw" podUID="71637165-aaba-4a30-abbf-641b66f7ad23" containerName="registry-server" containerID="cri-o://758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7" gracePeriod=2 Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.731852 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.834917 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-utilities\") pod \"71637165-aaba-4a30-abbf-641b66f7ad23\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.835183 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-catalog-content\") pod \"71637165-aaba-4a30-abbf-641b66f7ad23\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.835262 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8tr4\" (UniqueName: \"kubernetes.io/projected/71637165-aaba-4a30-abbf-641b66f7ad23-kube-api-access-z8tr4\") pod \"71637165-aaba-4a30-abbf-641b66f7ad23\" (UID: \"71637165-aaba-4a30-abbf-641b66f7ad23\") " Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.836065 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-utilities" (OuterVolumeSpecName: "utilities") pod "71637165-aaba-4a30-abbf-641b66f7ad23" (UID: "71637165-aaba-4a30-abbf-641b66f7ad23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.845975 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71637165-aaba-4a30-abbf-641b66f7ad23-kube-api-access-z8tr4" (OuterVolumeSpecName: "kube-api-access-z8tr4") pod "71637165-aaba-4a30-abbf-641b66f7ad23" (UID: "71637165-aaba-4a30-abbf-641b66f7ad23"). InnerVolumeSpecName "kube-api-access-z8tr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.853732 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71637165-aaba-4a30-abbf-641b66f7ad23" (UID: "71637165-aaba-4a30-abbf-641b66f7ad23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.937360 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.937420 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71637165-aaba-4a30-abbf-641b66f7ad23-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:37 crc kubenswrapper[4770]: I1004 03:54:37.937443 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8tr4\" (UniqueName: \"kubernetes.io/projected/71637165-aaba-4a30-abbf-641b66f7ad23-kube-api-access-z8tr4\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.195873 4770 generic.go:334] "Generic (PLEG): container finished" podID="71637165-aaba-4a30-abbf-641b66f7ad23" containerID="758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7" exitCode=0 Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.195946 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29dsw" event={"ID":"71637165-aaba-4a30-abbf-641b66f7ad23","Type":"ContainerDied","Data":"758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7"} Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.196027 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29dsw" event={"ID":"71637165-aaba-4a30-abbf-641b66f7ad23","Type":"ContainerDied","Data":"830c321390890aa496ac01691784b8dbae6b174e44bf94fb8fb313451367ad23"} Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.195995 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29dsw" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.196057 4770 scope.go:117] "RemoveContainer" containerID="758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.223671 4770 scope.go:117] "RemoveContainer" containerID="5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.249794 4770 scope.go:117] "RemoveContainer" containerID="b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.252125 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29dsw"] Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.256907 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-29dsw"] Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.286220 4770 scope.go:117] "RemoveContainer" containerID="758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7" Oct 04 03:54:38 crc kubenswrapper[4770]: E1004 03:54:38.286890 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7\": container with ID starting with 758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7 not found: ID does not exist" containerID="758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.286928 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7"} err="failed to get container status \"758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7\": rpc error: code = NotFound desc = could not find container \"758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7\": container with ID starting with 758d5597573b251f8c88a97cb0d5a857c06c4a2e331bd8bce82bdb0bd112b1c7 not found: ID does not exist" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.286958 4770 scope.go:117] "RemoveContainer" containerID="5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784" Oct 04 03:54:38 crc kubenswrapper[4770]: E1004 03:54:38.287657 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784\": container with ID starting with 5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784 not found: ID does not exist" containerID="5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.287766 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784"} err="failed to get container status \"5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784\": rpc error: code = NotFound desc = could not find container \"5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784\": container with ID starting with 5b33083cb42f75aeab41d816ba4df3210a4d3411c86f87667bcd12b33ba83784 not found: ID does not exist" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.287837 4770 scope.go:117] "RemoveContainer" containerID="b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654" Oct 04 03:54:38 crc kubenswrapper[4770]: E1004 03:54:38.288470 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654\": container with ID starting with b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654 not found: ID does not exist" containerID="b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654" Oct 04 03:54:38 crc kubenswrapper[4770]: I1004 03:54:38.288554 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654"} err="failed to get container status \"b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654\": rpc error: code = NotFound desc = could not find container \"b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654\": container with ID starting with b9096a80ba2a31d84be750ffb0df6224aef054cb5c57d036c2d7d873ebc5b654 not found: ID does not exist" Oct 04 03:54:39 crc kubenswrapper[4770]: I1004 03:54:39.690996 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71637165-aaba-4a30-abbf-641b66f7ad23" path="/var/lib/kubelet/pods/71637165-aaba-4a30-abbf-641b66f7ad23/volumes" Oct 04 03:55:01 crc kubenswrapper[4770]: I1004 03:55:01.796529 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:55:01 crc kubenswrapper[4770]: I1004 03:55:01.797253 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:55:31 crc kubenswrapper[4770]: I1004 03:55:31.795799 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:55:31 crc kubenswrapper[4770]: I1004 03:55:31.796470 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:55:31 crc kubenswrapper[4770]: I1004 03:55:31.796542 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 03:55:31 crc kubenswrapper[4770]: I1004 03:55:31.797453 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:55:31 crc kubenswrapper[4770]: I1004 03:55:31.797551 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" gracePeriod=600 Oct 04 03:55:31 crc kubenswrapper[4770]: E1004 03:55:31.939314 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:55:32 crc kubenswrapper[4770]: I1004 03:55:32.707566 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" exitCode=0 Oct 04 03:55:32 crc kubenswrapper[4770]: I1004 03:55:32.707655 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d"} Oct 04 03:55:32 crc kubenswrapper[4770]: I1004 03:55:32.707752 4770 scope.go:117] "RemoveContainer" containerID="eb8ca10a41c1b2139ab1071a91dbc916ec93f2edcb474db3d35896efc46be31b" Oct 04 03:55:32 crc kubenswrapper[4770]: I1004 03:55:32.708539 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:55:32 crc kubenswrapper[4770]: E1004 03:55:32.708929 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.255867 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2fxx2"] Oct 04 03:55:45 crc kubenswrapper[4770]: E1004 03:55:45.257681 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71637165-aaba-4a30-abbf-641b66f7ad23" containerName="registry-server" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.257717 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="71637165-aaba-4a30-abbf-641b66f7ad23" containerName="registry-server" Oct 04 03:55:45 crc kubenswrapper[4770]: E1004 03:55:45.257901 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71637165-aaba-4a30-abbf-641b66f7ad23" containerName="extract-content" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.257992 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="71637165-aaba-4a30-abbf-641b66f7ad23" containerName="extract-content" Oct 04 03:55:45 crc kubenswrapper[4770]: E1004 03:55:45.258151 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71637165-aaba-4a30-abbf-641b66f7ad23" containerName="extract-utilities" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.258197 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="71637165-aaba-4a30-abbf-641b66f7ad23" containerName="extract-utilities" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.258598 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="71637165-aaba-4a30-abbf-641b66f7ad23" containerName="registry-server" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.262147 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.270232 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2fxx2"] Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.359439 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77jqm\" (UniqueName: \"kubernetes.io/projected/26167cec-2f17-43a4-8395-a569920f4c8a-kube-api-access-77jqm\") pod \"redhat-operators-2fxx2\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.359504 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-utilities\") pod \"redhat-operators-2fxx2\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.359686 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-catalog-content\") pod \"redhat-operators-2fxx2\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.460685 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-catalog-content\") pod \"redhat-operators-2fxx2\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.460797 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77jqm\" (UniqueName: \"kubernetes.io/projected/26167cec-2f17-43a4-8395-a569920f4c8a-kube-api-access-77jqm\") pod \"redhat-operators-2fxx2\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.460843 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-utilities\") pod \"redhat-operators-2fxx2\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.461580 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-utilities\") pod \"redhat-operators-2fxx2\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.463485 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-catalog-content\") pod \"redhat-operators-2fxx2\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.493473 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77jqm\" (UniqueName: \"kubernetes.io/projected/26167cec-2f17-43a4-8395-a569920f4c8a-kube-api-access-77jqm\") pod \"redhat-operators-2fxx2\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.610091 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:45 crc kubenswrapper[4770]: I1004 03:55:45.855880 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2fxx2"] Oct 04 03:55:46 crc kubenswrapper[4770]: I1004 03:55:46.673660 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:55:46 crc kubenswrapper[4770]: E1004 03:55:46.674071 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:55:46 crc kubenswrapper[4770]: I1004 03:55:46.860275 4770 generic.go:334] "Generic (PLEG): container finished" podID="26167cec-2f17-43a4-8395-a569920f4c8a" containerID="23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c" exitCode=0 Oct 04 03:55:46 crc kubenswrapper[4770]: I1004 03:55:46.860489 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fxx2" event={"ID":"26167cec-2f17-43a4-8395-a569920f4c8a","Type":"ContainerDied","Data":"23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c"} Oct 04 03:55:46 crc kubenswrapper[4770]: I1004 03:55:46.860814 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fxx2" event={"ID":"26167cec-2f17-43a4-8395-a569920f4c8a","Type":"ContainerStarted","Data":"c335030dfdbf48e9a75034c9d425f144d783c214eea61cc16130169bd5fe04fd"} Oct 04 03:55:48 crc kubenswrapper[4770]: I1004 03:55:48.882501 4770 generic.go:334] "Generic (PLEG): container finished" podID="26167cec-2f17-43a4-8395-a569920f4c8a" containerID="7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158" exitCode=0 Oct 04 03:55:48 crc kubenswrapper[4770]: I1004 03:55:48.882605 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fxx2" event={"ID":"26167cec-2f17-43a4-8395-a569920f4c8a","Type":"ContainerDied","Data":"7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158"} Oct 04 03:55:49 crc kubenswrapper[4770]: I1004 03:55:49.898409 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fxx2" event={"ID":"26167cec-2f17-43a4-8395-a569920f4c8a","Type":"ContainerStarted","Data":"bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48"} Oct 04 03:55:49 crc kubenswrapper[4770]: I1004 03:55:49.939075 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2fxx2" podStartSLOduration=2.471883417 podStartE2EDuration="4.939050815s" podCreationTimestamp="2025-10-04 03:55:45 +0000 UTC" firstStartedPulling="2025-10-04 03:55:46.863535103 +0000 UTC m=+3158.155544825" lastFinishedPulling="2025-10-04 03:55:49.330702471 +0000 UTC m=+3160.622712223" observedRunningTime="2025-10-04 03:55:49.930332917 +0000 UTC m=+3161.222342639" watchObservedRunningTime="2025-10-04 03:55:49.939050815 +0000 UTC m=+3161.231060557" Oct 04 03:55:55 crc kubenswrapper[4770]: I1004 03:55:55.610476 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:55 crc kubenswrapper[4770]: I1004 03:55:55.611090 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:55 crc kubenswrapper[4770]: I1004 03:55:55.687074 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:56 crc kubenswrapper[4770]: I1004 03:55:56.036760 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:56 crc kubenswrapper[4770]: I1004 03:55:56.107092 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2fxx2"] Oct 04 03:55:57 crc kubenswrapper[4770]: I1004 03:55:57.975237 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2fxx2" podUID="26167cec-2f17-43a4-8395-a569920f4c8a" containerName="registry-server" containerID="cri-o://bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48" gracePeriod=2 Oct 04 03:55:58 crc kubenswrapper[4770]: I1004 03:55:58.966920 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:58 crc kubenswrapper[4770]: I1004 03:55:58.988210 4770 generic.go:334] "Generic (PLEG): container finished" podID="26167cec-2f17-43a4-8395-a569920f4c8a" containerID="bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48" exitCode=0 Oct 04 03:55:58 crc kubenswrapper[4770]: I1004 03:55:58.988256 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fxx2" event={"ID":"26167cec-2f17-43a4-8395-a569920f4c8a","Type":"ContainerDied","Data":"bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48"} Oct 04 03:55:58 crc kubenswrapper[4770]: I1004 03:55:58.988291 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fxx2" event={"ID":"26167cec-2f17-43a4-8395-a569920f4c8a","Type":"ContainerDied","Data":"c335030dfdbf48e9a75034c9d425f144d783c214eea61cc16130169bd5fe04fd"} Oct 04 03:55:58 crc kubenswrapper[4770]: I1004 03:55:58.988314 4770 scope.go:117] "RemoveContainer" containerID="bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48" Oct 04 03:55:58 crc kubenswrapper[4770]: I1004 03:55:58.988589 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2fxx2" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.030535 4770 scope.go:117] "RemoveContainer" containerID="7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.053297 4770 scope.go:117] "RemoveContainer" containerID="23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.065842 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-utilities\") pod \"26167cec-2f17-43a4-8395-a569920f4c8a\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.066149 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77jqm\" (UniqueName: \"kubernetes.io/projected/26167cec-2f17-43a4-8395-a569920f4c8a-kube-api-access-77jqm\") pod \"26167cec-2f17-43a4-8395-a569920f4c8a\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.066277 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-catalog-content\") pod \"26167cec-2f17-43a4-8395-a569920f4c8a\" (UID: \"26167cec-2f17-43a4-8395-a569920f4c8a\") " Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.067729 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-utilities" (OuterVolumeSpecName: "utilities") pod "26167cec-2f17-43a4-8395-a569920f4c8a" (UID: "26167cec-2f17-43a4-8395-a569920f4c8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.068459 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.076970 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26167cec-2f17-43a4-8395-a569920f4c8a-kube-api-access-77jqm" (OuterVolumeSpecName: "kube-api-access-77jqm") pod "26167cec-2f17-43a4-8395-a569920f4c8a" (UID: "26167cec-2f17-43a4-8395-a569920f4c8a"). InnerVolumeSpecName "kube-api-access-77jqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.079391 4770 scope.go:117] "RemoveContainer" containerID="bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48" Oct 04 03:55:59 crc kubenswrapper[4770]: E1004 03:55:59.079989 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48\": container with ID starting with bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48 not found: ID does not exist" containerID="bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.080039 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48"} err="failed to get container status \"bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48\": rpc error: code = NotFound desc = could not find container \"bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48\": container with ID starting with bee47e7b3d7fd0c760744b3abddc14a639a53d74ff75e4bfe7682cfbbd6cfb48 not found: ID does not exist" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.080070 4770 scope.go:117] "RemoveContainer" containerID="7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158" Oct 04 03:55:59 crc kubenswrapper[4770]: E1004 03:55:59.080457 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158\": container with ID starting with 7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158 not found: ID does not exist" containerID="7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.080492 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158"} err="failed to get container status \"7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158\": rpc error: code = NotFound desc = could not find container \"7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158\": container with ID starting with 7c8d6d178c73c9f8c62ca6fed19e518c6638b4c2cb9da105c7420454ae482158 not found: ID does not exist" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.080525 4770 scope.go:117] "RemoveContainer" containerID="23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c" Oct 04 03:55:59 crc kubenswrapper[4770]: E1004 03:55:59.080860 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c\": container with ID starting with 23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c not found: ID does not exist" containerID="23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.080878 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c"} err="failed to get container status \"23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c\": rpc error: code = NotFound desc = could not find container \"23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c\": container with ID starting with 23855be69678d7343ba6615cc49ec694cbff4f5e8ad6a61bbcbaae0963e1724c not found: ID does not exist" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.170306 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77jqm\" (UniqueName: \"kubernetes.io/projected/26167cec-2f17-43a4-8395-a569920f4c8a-kube-api-access-77jqm\") on node \"crc\" DevicePath \"\"" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.805772 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26167cec-2f17-43a4-8395-a569920f4c8a" (UID: "26167cec-2f17-43a4-8395-a569920f4c8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.883292 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26167cec-2f17-43a4-8395-a569920f4c8a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.942503 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2fxx2"] Oct 04 03:55:59 crc kubenswrapper[4770]: I1004 03:55:59.964645 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2fxx2"] Oct 04 03:56:01 crc kubenswrapper[4770]: I1004 03:56:01.674070 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:56:01 crc kubenswrapper[4770]: E1004 03:56:01.674974 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:56:01 crc kubenswrapper[4770]: I1004 03:56:01.690553 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26167cec-2f17-43a4-8395-a569920f4c8a" path="/var/lib/kubelet/pods/26167cec-2f17-43a4-8395-a569920f4c8a/volumes" Oct 04 03:56:12 crc kubenswrapper[4770]: I1004 03:56:12.673537 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:56:12 crc kubenswrapper[4770]: E1004 03:56:12.674209 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:56:27 crc kubenswrapper[4770]: I1004 03:56:27.674457 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:56:27 crc kubenswrapper[4770]: E1004 03:56:27.675701 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:56:40 crc kubenswrapper[4770]: I1004 03:56:40.673640 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:56:40 crc kubenswrapper[4770]: E1004 03:56:40.674773 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:56:51 crc kubenswrapper[4770]: I1004 03:56:51.674086 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:56:51 crc kubenswrapper[4770]: E1004 03:56:51.674986 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:57:02 crc kubenswrapper[4770]: I1004 03:57:02.674062 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:57:02 crc kubenswrapper[4770]: E1004 03:57:02.675044 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:57:14 crc kubenswrapper[4770]: I1004 03:57:14.674264 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:57:14 crc kubenswrapper[4770]: E1004 03:57:14.675222 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:57:26 crc kubenswrapper[4770]: I1004 03:57:26.673447 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:57:26 crc kubenswrapper[4770]: E1004 03:57:26.674207 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:57:37 crc kubenswrapper[4770]: I1004 03:57:37.674572 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:57:37 crc kubenswrapper[4770]: E1004 03:57:37.675421 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:57:51 crc kubenswrapper[4770]: I1004 03:57:51.673904 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:57:51 crc kubenswrapper[4770]: E1004 03:57:51.676418 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:58:06 crc kubenswrapper[4770]: I1004 03:58:06.674371 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:58:06 crc kubenswrapper[4770]: E1004 03:58:06.675344 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:58:20 crc kubenswrapper[4770]: I1004 03:58:20.673853 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:58:20 crc kubenswrapper[4770]: E1004 03:58:20.674833 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:58:33 crc kubenswrapper[4770]: I1004 03:58:33.674132 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:58:33 crc kubenswrapper[4770]: E1004 03:58:33.675053 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:58:47 crc kubenswrapper[4770]: I1004 03:58:47.673655 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:58:47 crc kubenswrapper[4770]: E1004 03:58:47.674364 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:59:01 crc kubenswrapper[4770]: I1004 03:59:01.673879 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:59:01 crc kubenswrapper[4770]: E1004 03:59:01.675254 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:59:16 crc kubenswrapper[4770]: I1004 03:59:16.674438 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:59:16 crc kubenswrapper[4770]: E1004 03:59:16.675522 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:59:27 crc kubenswrapper[4770]: I1004 03:59:27.674396 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:59:27 crc kubenswrapper[4770]: E1004 03:59:27.675745 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.090516 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l98fz"] Oct 04 03:59:33 crc kubenswrapper[4770]: E1004 03:59:33.092042 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26167cec-2f17-43a4-8395-a569920f4c8a" containerName="extract-content" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.092135 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="26167cec-2f17-43a4-8395-a569920f4c8a" containerName="extract-content" Oct 04 03:59:33 crc kubenswrapper[4770]: E1004 03:59:33.092198 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26167cec-2f17-43a4-8395-a569920f4c8a" containerName="registry-server" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.092251 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="26167cec-2f17-43a4-8395-a569920f4c8a" containerName="registry-server" Oct 04 03:59:33 crc kubenswrapper[4770]: E1004 03:59:33.092321 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26167cec-2f17-43a4-8395-a569920f4c8a" containerName="extract-utilities" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.092378 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="26167cec-2f17-43a4-8395-a569920f4c8a" containerName="extract-utilities" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.092572 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="26167cec-2f17-43a4-8395-a569920f4c8a" containerName="registry-server" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.093590 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.099096 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l98fz"] Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.176452 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-utilities\") pod \"certified-operators-l98fz\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.176697 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-catalog-content\") pod \"certified-operators-l98fz\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.176879 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnlgm\" (UniqueName: \"kubernetes.io/projected/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-kube-api-access-pnlgm\") pod \"certified-operators-l98fz\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.279156 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-utilities\") pod \"certified-operators-l98fz\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.279238 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-catalog-content\") pod \"certified-operators-l98fz\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.279281 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnlgm\" (UniqueName: \"kubernetes.io/projected/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-kube-api-access-pnlgm\") pod \"certified-operators-l98fz\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.280141 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-utilities\") pod \"certified-operators-l98fz\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.280361 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-catalog-content\") pod \"certified-operators-l98fz\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.313453 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnlgm\" (UniqueName: \"kubernetes.io/projected/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-kube-api-access-pnlgm\") pod \"certified-operators-l98fz\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.452437 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:33 crc kubenswrapper[4770]: I1004 03:59:33.906787 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l98fz"] Oct 04 03:59:33 crc kubenswrapper[4770]: W1004 03:59:33.918597 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f039fd5_f8d9_4c02_9d8c_ea0b5891907e.slice/crio-666a2de6191b004d5ff73cca4c639c7fd5c49c80c1d86e711e219394fbb1a53e WatchSource:0}: Error finding container 666a2de6191b004d5ff73cca4c639c7fd5c49c80c1d86e711e219394fbb1a53e: Status 404 returned error can't find the container with id 666a2de6191b004d5ff73cca4c639c7fd5c49c80c1d86e711e219394fbb1a53e Oct 04 03:59:34 crc kubenswrapper[4770]: I1004 03:59:34.939056 4770 generic.go:334] "Generic (PLEG): container finished" podID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerID="2920490213ad32259f3fb381e76215116e532576711401b7fce1463ede896b1f" exitCode=0 Oct 04 03:59:34 crc kubenswrapper[4770]: I1004 03:59:34.939134 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l98fz" event={"ID":"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e","Type":"ContainerDied","Data":"2920490213ad32259f3fb381e76215116e532576711401b7fce1463ede896b1f"} Oct 04 03:59:34 crc kubenswrapper[4770]: I1004 03:59:34.939436 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l98fz" event={"ID":"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e","Type":"ContainerStarted","Data":"666a2de6191b004d5ff73cca4c639c7fd5c49c80c1d86e711e219394fbb1a53e"} Oct 04 03:59:34 crc kubenswrapper[4770]: I1004 03:59:34.943558 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:59:35 crc kubenswrapper[4770]: I1004 03:59:35.948787 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l98fz" event={"ID":"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e","Type":"ContainerStarted","Data":"46ee569fc0a136f604c0f6da5a3f6cb30713b3ab95e6fae5de6743ee5f5f8d31"} Oct 04 03:59:36 crc kubenswrapper[4770]: I1004 03:59:36.962640 4770 generic.go:334] "Generic (PLEG): container finished" podID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerID="46ee569fc0a136f604c0f6da5a3f6cb30713b3ab95e6fae5de6743ee5f5f8d31" exitCode=0 Oct 04 03:59:36 crc kubenswrapper[4770]: I1004 03:59:36.962699 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l98fz" event={"ID":"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e","Type":"ContainerDied","Data":"46ee569fc0a136f604c0f6da5a3f6cb30713b3ab95e6fae5de6743ee5f5f8d31"} Oct 04 03:59:38 crc kubenswrapper[4770]: I1004 03:59:38.982440 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l98fz" event={"ID":"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e","Type":"ContainerStarted","Data":"824c8182bbdf20638d36e3d1aacc1bfce751ebeb78beb3ae14c42a06076b0730"} Oct 04 03:59:39 crc kubenswrapper[4770]: I1004 03:59:39.007327 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l98fz" podStartSLOduration=3.025029707 podStartE2EDuration="6.007305835s" podCreationTimestamp="2025-10-04 03:59:33 +0000 UTC" firstStartedPulling="2025-10-04 03:59:34.94290461 +0000 UTC m=+3386.234914362" lastFinishedPulling="2025-10-04 03:59:37.925180788 +0000 UTC m=+3389.217190490" observedRunningTime="2025-10-04 03:59:39.006640177 +0000 UTC m=+3390.298649939" watchObservedRunningTime="2025-10-04 03:59:39.007305835 +0000 UTC m=+3390.299315557" Oct 04 03:59:39 crc kubenswrapper[4770]: I1004 03:59:39.683510 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:59:39 crc kubenswrapper[4770]: E1004 03:59:39.683985 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 03:59:43 crc kubenswrapper[4770]: I1004 03:59:43.452603 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:43 crc kubenswrapper[4770]: I1004 03:59:43.453520 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:43 crc kubenswrapper[4770]: I1004 03:59:43.529469 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:44 crc kubenswrapper[4770]: I1004 03:59:44.107832 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:44 crc kubenswrapper[4770]: I1004 03:59:44.178232 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l98fz"] Oct 04 03:59:46 crc kubenswrapper[4770]: I1004 03:59:46.054693 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l98fz" podUID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerName="registry-server" containerID="cri-o://824c8182bbdf20638d36e3d1aacc1bfce751ebeb78beb3ae14c42a06076b0730" gracePeriod=2 Oct 04 03:59:47 crc kubenswrapper[4770]: I1004 03:59:47.067526 4770 generic.go:334] "Generic (PLEG): container finished" podID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerID="824c8182bbdf20638d36e3d1aacc1bfce751ebeb78beb3ae14c42a06076b0730" exitCode=0 Oct 04 03:59:47 crc kubenswrapper[4770]: I1004 03:59:47.067615 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l98fz" event={"ID":"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e","Type":"ContainerDied","Data":"824c8182bbdf20638d36e3d1aacc1bfce751ebeb78beb3ae14c42a06076b0730"} Oct 04 03:59:48 crc kubenswrapper[4770]: I1004 03:59:48.403859 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:48 crc kubenswrapper[4770]: I1004 03:59:48.526484 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-utilities\") pod \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " Oct 04 03:59:48 crc kubenswrapper[4770]: I1004 03:59:48.526536 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-catalog-content\") pod \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " Oct 04 03:59:48 crc kubenswrapper[4770]: I1004 03:59:48.526621 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnlgm\" (UniqueName: \"kubernetes.io/projected/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-kube-api-access-pnlgm\") pod \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\" (UID: \"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e\") " Oct 04 03:59:48 crc kubenswrapper[4770]: I1004 03:59:48.528143 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-utilities" (OuterVolumeSpecName: "utilities") pod "8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" (UID: "8f039fd5-f8d9-4c02-9d8c-ea0b5891907e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:59:48 crc kubenswrapper[4770]: I1004 03:59:48.534251 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-kube-api-access-pnlgm" (OuterVolumeSpecName: "kube-api-access-pnlgm") pod "8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" (UID: "8f039fd5-f8d9-4c02-9d8c-ea0b5891907e"). InnerVolumeSpecName "kube-api-access-pnlgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:48 crc kubenswrapper[4770]: I1004 03:59:48.628381 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnlgm\" (UniqueName: \"kubernetes.io/projected/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-kube-api-access-pnlgm\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:48 crc kubenswrapper[4770]: I1004 03:59:48.628430 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.092375 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l98fz" event={"ID":"8f039fd5-f8d9-4c02-9d8c-ea0b5891907e","Type":"ContainerDied","Data":"666a2de6191b004d5ff73cca4c639c7fd5c49c80c1d86e711e219394fbb1a53e"} Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.092460 4770 scope.go:117] "RemoveContainer" containerID="824c8182bbdf20638d36e3d1aacc1bfce751ebeb78beb3ae14c42a06076b0730" Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.092483 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l98fz" Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.124235 4770 scope.go:117] "RemoveContainer" containerID="46ee569fc0a136f604c0f6da5a3f6cb30713b3ab95e6fae5de6743ee5f5f8d31" Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.155645 4770 scope.go:117] "RemoveContainer" containerID="2920490213ad32259f3fb381e76215116e532576711401b7fce1463ede896b1f" Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.199136 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" (UID: "8f039fd5-f8d9-4c02-9d8c-ea0b5891907e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.240077 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.450625 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l98fz"] Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.457150 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l98fz"] Oct 04 03:59:49 crc kubenswrapper[4770]: I1004 03:59:49.688998 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" path="/var/lib/kubelet/pods/8f039fd5-f8d9-4c02-9d8c-ea0b5891907e/volumes" Oct 04 03:59:51 crc kubenswrapper[4770]: I1004 03:59:51.674559 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 03:59:51 crc kubenswrapper[4770]: E1004 03:59:51.675498 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.213275 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4"] Oct 04 04:00:00 crc kubenswrapper[4770]: E1004 04:00:00.214604 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerName="extract-content" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.214635 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerName="extract-content" Oct 04 04:00:00 crc kubenswrapper[4770]: E1004 04:00:00.214679 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerName="extract-utilities" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.214697 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerName="extract-utilities" Oct 04 04:00:00 crc kubenswrapper[4770]: E1004 04:00:00.214730 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerName="registry-server" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.214747 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerName="registry-server" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.215498 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f039fd5-f8d9-4c02-9d8c-ea0b5891907e" containerName="registry-server" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.216532 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.220125 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.220694 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.225889 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4"] Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.317975 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3463bd44-4352-4f30-9ebd-ac8667f4db7f-secret-volume\") pod \"collect-profiles-29325840-t68f4\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.318140 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3463bd44-4352-4f30-9ebd-ac8667f4db7f-config-volume\") pod \"collect-profiles-29325840-t68f4\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.318182 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxs6v\" (UniqueName: \"kubernetes.io/projected/3463bd44-4352-4f30-9ebd-ac8667f4db7f-kube-api-access-dxs6v\") pod \"collect-profiles-29325840-t68f4\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.419090 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxs6v\" (UniqueName: \"kubernetes.io/projected/3463bd44-4352-4f30-9ebd-ac8667f4db7f-kube-api-access-dxs6v\") pod \"collect-profiles-29325840-t68f4\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.419501 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3463bd44-4352-4f30-9ebd-ac8667f4db7f-secret-volume\") pod \"collect-profiles-29325840-t68f4\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.419895 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3463bd44-4352-4f30-9ebd-ac8667f4db7f-config-volume\") pod \"collect-profiles-29325840-t68f4\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.421586 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3463bd44-4352-4f30-9ebd-ac8667f4db7f-config-volume\") pod \"collect-profiles-29325840-t68f4\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.426748 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3463bd44-4352-4f30-9ebd-ac8667f4db7f-secret-volume\") pod \"collect-profiles-29325840-t68f4\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.442221 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxs6v\" (UniqueName: \"kubernetes.io/projected/3463bd44-4352-4f30-9ebd-ac8667f4db7f-kube-api-access-dxs6v\") pod \"collect-profiles-29325840-t68f4\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:00 crc kubenswrapper[4770]: I1004 04:00:00.550382 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:01 crc kubenswrapper[4770]: I1004 04:00:01.018802 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4"] Oct 04 04:00:01 crc kubenswrapper[4770]: W1004 04:00:01.030970 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3463bd44_4352_4f30_9ebd_ac8667f4db7f.slice/crio-3ebbb78e27935031be091f1387fc24e20ec5d7d1a9b832b82f2fab5de92af955 WatchSource:0}: Error finding container 3ebbb78e27935031be091f1387fc24e20ec5d7d1a9b832b82f2fab5de92af955: Status 404 returned error can't find the container with id 3ebbb78e27935031be091f1387fc24e20ec5d7d1a9b832b82f2fab5de92af955 Oct 04 04:00:01 crc kubenswrapper[4770]: I1004 04:00:01.239139 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" event={"ID":"3463bd44-4352-4f30-9ebd-ac8667f4db7f","Type":"ContainerStarted","Data":"ff3bd21e1fd61fb6741aa77b9ef6bf4e849ceb0e4089b0a8419e3d8671ca3f43"} Oct 04 04:00:01 crc kubenswrapper[4770]: I1004 04:00:01.239522 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" event={"ID":"3463bd44-4352-4f30-9ebd-ac8667f4db7f","Type":"ContainerStarted","Data":"3ebbb78e27935031be091f1387fc24e20ec5d7d1a9b832b82f2fab5de92af955"} Oct 04 04:00:01 crc kubenswrapper[4770]: I1004 04:00:01.262453 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" podStartSLOduration=1.2624334990000001 podStartE2EDuration="1.262433499s" podCreationTimestamp="2025-10-04 04:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:01.25862824 +0000 UTC m=+3412.550637962" watchObservedRunningTime="2025-10-04 04:00:01.262433499 +0000 UTC m=+3412.554443221" Oct 04 04:00:02 crc kubenswrapper[4770]: I1004 04:00:02.250434 4770 generic.go:334] "Generic (PLEG): container finished" podID="3463bd44-4352-4f30-9ebd-ac8667f4db7f" containerID="ff3bd21e1fd61fb6741aa77b9ef6bf4e849ceb0e4089b0a8419e3d8671ca3f43" exitCode=0 Oct 04 04:00:02 crc kubenswrapper[4770]: I1004 04:00:02.250487 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" event={"ID":"3463bd44-4352-4f30-9ebd-ac8667f4db7f","Type":"ContainerDied","Data":"ff3bd21e1fd61fb6741aa77b9ef6bf4e849ceb0e4089b0a8419e3d8671ca3f43"} Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.594635 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.670034 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3463bd44-4352-4f30-9ebd-ac8667f4db7f-secret-volume\") pod \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.670087 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxs6v\" (UniqueName: \"kubernetes.io/projected/3463bd44-4352-4f30-9ebd-ac8667f4db7f-kube-api-access-dxs6v\") pod \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.670131 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3463bd44-4352-4f30-9ebd-ac8667f4db7f-config-volume\") pod \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\" (UID: \"3463bd44-4352-4f30-9ebd-ac8667f4db7f\") " Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.671161 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3463bd44-4352-4f30-9ebd-ac8667f4db7f-config-volume" (OuterVolumeSpecName: "config-volume") pod "3463bd44-4352-4f30-9ebd-ac8667f4db7f" (UID: "3463bd44-4352-4f30-9ebd-ac8667f4db7f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.673863 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 04:00:03 crc kubenswrapper[4770]: E1004 04:00:03.674273 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.677763 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3463bd44-4352-4f30-9ebd-ac8667f4db7f-kube-api-access-dxs6v" (OuterVolumeSpecName: "kube-api-access-dxs6v") pod "3463bd44-4352-4f30-9ebd-ac8667f4db7f" (UID: "3463bd44-4352-4f30-9ebd-ac8667f4db7f"). InnerVolumeSpecName "kube-api-access-dxs6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.679302 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3463bd44-4352-4f30-9ebd-ac8667f4db7f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3463bd44-4352-4f30-9ebd-ac8667f4db7f" (UID: "3463bd44-4352-4f30-9ebd-ac8667f4db7f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.771743 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxs6v\" (UniqueName: \"kubernetes.io/projected/3463bd44-4352-4f30-9ebd-ac8667f4db7f-kube-api-access-dxs6v\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.771811 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3463bd44-4352-4f30-9ebd-ac8667f4db7f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:03 crc kubenswrapper[4770]: I1004 04:00:03.771837 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3463bd44-4352-4f30-9ebd-ac8667f4db7f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:04 crc kubenswrapper[4770]: I1004 04:00:04.270681 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" event={"ID":"3463bd44-4352-4f30-9ebd-ac8667f4db7f","Type":"ContainerDied","Data":"3ebbb78e27935031be091f1387fc24e20ec5d7d1a9b832b82f2fab5de92af955"} Oct 04 04:00:04 crc kubenswrapper[4770]: I1004 04:00:04.270738 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ebbb78e27935031be091f1387fc24e20ec5d7d1a9b832b82f2fab5de92af955" Oct 04 04:00:04 crc kubenswrapper[4770]: I1004 04:00:04.270861 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4" Oct 04 04:00:04 crc kubenswrapper[4770]: I1004 04:00:04.371834 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w"] Oct 04 04:00:04 crc kubenswrapper[4770]: I1004 04:00:04.376553 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325795-hrw4w"] Oct 04 04:00:05 crc kubenswrapper[4770]: I1004 04:00:05.691457 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6739a0a5-d8da-49e3-8131-d2e9607a0455" path="/var/lib/kubelet/pods/6739a0a5-d8da-49e3-8131-d2e9607a0455/volumes" Oct 04 04:00:14 crc kubenswrapper[4770]: I1004 04:00:14.674455 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 04:00:14 crc kubenswrapper[4770]: E1004 04:00:14.677699 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:00:23 crc kubenswrapper[4770]: I1004 04:00:23.416953 4770 scope.go:117] "RemoveContainer" containerID="4bd7c972b5ebb9e930952a683340a6cad67fb4154f04dd3135db420e4dc96a73" Oct 04 04:00:26 crc kubenswrapper[4770]: I1004 04:00:26.674416 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 04:00:26 crc kubenswrapper[4770]: E1004 04:00:26.675335 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:00:38 crc kubenswrapper[4770]: I1004 04:00:38.673808 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 04:00:39 crc kubenswrapper[4770]: I1004 04:00:39.627053 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"ef02c7e0d8b738ce4aa4b7ac3b28fc6a0bf21d37757a1d3f1e1ec6426badd23a"} Oct 04 04:03:01 crc kubenswrapper[4770]: I1004 04:03:01.795967 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:03:01 crc kubenswrapper[4770]: I1004 04:03:01.796575 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:03:31 crc kubenswrapper[4770]: I1004 04:03:31.795507 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:03:31 crc kubenswrapper[4770]: I1004 04:03:31.796419 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:04:01 crc kubenswrapper[4770]: I1004 04:04:01.796036 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:04:01 crc kubenswrapper[4770]: I1004 04:04:01.796600 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:04:01 crc kubenswrapper[4770]: I1004 04:04:01.796667 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:04:01 crc kubenswrapper[4770]: I1004 04:04:01.797615 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ef02c7e0d8b738ce4aa4b7ac3b28fc6a0bf21d37757a1d3f1e1ec6426badd23a"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:04:01 crc kubenswrapper[4770]: I1004 04:04:01.797731 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://ef02c7e0d8b738ce4aa4b7ac3b28fc6a0bf21d37757a1d3f1e1ec6426badd23a" gracePeriod=600 Oct 04 04:04:02 crc kubenswrapper[4770]: I1004 04:04:02.617024 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="ef02c7e0d8b738ce4aa4b7ac3b28fc6a0bf21d37757a1d3f1e1ec6426badd23a" exitCode=0 Oct 04 04:04:02 crc kubenswrapper[4770]: I1004 04:04:02.617105 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"ef02c7e0d8b738ce4aa4b7ac3b28fc6a0bf21d37757a1d3f1e1ec6426badd23a"} Oct 04 04:04:02 crc kubenswrapper[4770]: I1004 04:04:02.617335 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258"} Oct 04 04:04:02 crc kubenswrapper[4770]: I1004 04:04:02.617355 4770 scope.go:117] "RemoveContainer" containerID="53f39520fcd67a3471ed633dec681366e4f0ec10c46eb0bb07c3274c5a05c60d" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.402033 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4srnq"] Oct 04 04:04:36 crc kubenswrapper[4770]: E1004 04:04:36.403611 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3463bd44-4352-4f30-9ebd-ac8667f4db7f" containerName="collect-profiles" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.403638 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3463bd44-4352-4f30-9ebd-ac8667f4db7f" containerName="collect-profiles" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.403882 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3463bd44-4352-4f30-9ebd-ac8667f4db7f" containerName="collect-profiles" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.405660 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.406929 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzj8d\" (UniqueName: \"kubernetes.io/projected/b50ecf50-3ddc-416e-9838-06d036a68be5-kube-api-access-fzj8d\") pod \"redhat-marketplace-4srnq\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.407216 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-utilities\") pod \"redhat-marketplace-4srnq\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.407280 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-catalog-content\") pod \"redhat-marketplace-4srnq\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.424882 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4srnq"] Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.508569 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-utilities\") pod \"redhat-marketplace-4srnq\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.508849 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-catalog-content\") pod \"redhat-marketplace-4srnq\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.509130 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzj8d\" (UniqueName: \"kubernetes.io/projected/b50ecf50-3ddc-416e-9838-06d036a68be5-kube-api-access-fzj8d\") pod \"redhat-marketplace-4srnq\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.509278 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-utilities\") pod \"redhat-marketplace-4srnq\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.509302 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-catalog-content\") pod \"redhat-marketplace-4srnq\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.538317 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzj8d\" (UniqueName: \"kubernetes.io/projected/b50ecf50-3ddc-416e-9838-06d036a68be5-kube-api-access-fzj8d\") pod \"redhat-marketplace-4srnq\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:36 crc kubenswrapper[4770]: I1004 04:04:36.754749 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:37 crc kubenswrapper[4770]: I1004 04:04:37.208102 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4srnq"] Oct 04 04:04:37 crc kubenswrapper[4770]: I1004 04:04:37.968678 4770 generic.go:334] "Generic (PLEG): container finished" podID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerID="0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62" exitCode=0 Oct 04 04:04:37 crc kubenswrapper[4770]: I1004 04:04:37.968728 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4srnq" event={"ID":"b50ecf50-3ddc-416e-9838-06d036a68be5","Type":"ContainerDied","Data":"0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62"} Oct 04 04:04:37 crc kubenswrapper[4770]: I1004 04:04:37.969057 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4srnq" event={"ID":"b50ecf50-3ddc-416e-9838-06d036a68be5","Type":"ContainerStarted","Data":"72fee1e259821c36fe1320d1097d2c65faba11d2c2c2d9bfec684007a9b970cb"} Oct 04 04:04:37 crc kubenswrapper[4770]: I1004 04:04:37.971662 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:04:38 crc kubenswrapper[4770]: I1004 04:04:38.983356 4770 generic.go:334] "Generic (PLEG): container finished" podID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerID="dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706" exitCode=0 Oct 04 04:04:38 crc kubenswrapper[4770]: I1004 04:04:38.983523 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4srnq" event={"ID":"b50ecf50-3ddc-416e-9838-06d036a68be5","Type":"ContainerDied","Data":"dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706"} Oct 04 04:04:39 crc kubenswrapper[4770]: I1004 04:04:39.996378 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4srnq" event={"ID":"b50ecf50-3ddc-416e-9838-06d036a68be5","Type":"ContainerStarted","Data":"8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324"} Oct 04 04:04:40 crc kubenswrapper[4770]: I1004 04:04:40.024847 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4srnq" podStartSLOduration=2.616502189 podStartE2EDuration="4.024828892s" podCreationTimestamp="2025-10-04 04:04:36 +0000 UTC" firstStartedPulling="2025-10-04 04:04:37.971438503 +0000 UTC m=+3689.263448215" lastFinishedPulling="2025-10-04 04:04:39.379765206 +0000 UTC m=+3690.671774918" observedRunningTime="2025-10-04 04:04:40.018913448 +0000 UTC m=+3691.310923180" watchObservedRunningTime="2025-10-04 04:04:40.024828892 +0000 UTC m=+3691.316838614" Oct 04 04:04:46 crc kubenswrapper[4770]: I1004 04:04:46.755551 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:46 crc kubenswrapper[4770]: I1004 04:04:46.756426 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:46 crc kubenswrapper[4770]: I1004 04:04:46.831791 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:47 crc kubenswrapper[4770]: I1004 04:04:47.128291 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:47 crc kubenswrapper[4770]: I1004 04:04:47.189555 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4srnq"] Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.078723 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4srnq" podUID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerName="registry-server" containerID="cri-o://8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324" gracePeriod=2 Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.477511 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.612800 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-catalog-content\") pod \"b50ecf50-3ddc-416e-9838-06d036a68be5\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.612951 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzj8d\" (UniqueName: \"kubernetes.io/projected/b50ecf50-3ddc-416e-9838-06d036a68be5-kube-api-access-fzj8d\") pod \"b50ecf50-3ddc-416e-9838-06d036a68be5\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.613177 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-utilities\") pod \"b50ecf50-3ddc-416e-9838-06d036a68be5\" (UID: \"b50ecf50-3ddc-416e-9838-06d036a68be5\") " Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.615743 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-utilities" (OuterVolumeSpecName: "utilities") pod "b50ecf50-3ddc-416e-9838-06d036a68be5" (UID: "b50ecf50-3ddc-416e-9838-06d036a68be5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.635752 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b50ecf50-3ddc-416e-9838-06d036a68be5-kube-api-access-fzj8d" (OuterVolumeSpecName: "kube-api-access-fzj8d") pod "b50ecf50-3ddc-416e-9838-06d036a68be5" (UID: "b50ecf50-3ddc-416e-9838-06d036a68be5"). InnerVolumeSpecName "kube-api-access-fzj8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.638087 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b50ecf50-3ddc-416e-9838-06d036a68be5" (UID: "b50ecf50-3ddc-416e-9838-06d036a68be5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.716185 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.716221 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzj8d\" (UniqueName: \"kubernetes.io/projected/b50ecf50-3ddc-416e-9838-06d036a68be5-kube-api-access-fzj8d\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:49 crc kubenswrapper[4770]: I1004 04:04:49.716238 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50ecf50-3ddc-416e-9838-06d036a68be5-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.092552 4770 generic.go:334] "Generic (PLEG): container finished" podID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerID="8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324" exitCode=0 Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.092643 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4srnq" event={"ID":"b50ecf50-3ddc-416e-9838-06d036a68be5","Type":"ContainerDied","Data":"8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324"} Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.093119 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4srnq" event={"ID":"b50ecf50-3ddc-416e-9838-06d036a68be5","Type":"ContainerDied","Data":"72fee1e259821c36fe1320d1097d2c65faba11d2c2c2d9bfec684007a9b970cb"} Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.093155 4770 scope.go:117] "RemoveContainer" containerID="8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.092704 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4srnq" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.126901 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4srnq"] Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.130821 4770 scope.go:117] "RemoveContainer" containerID="dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.136635 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4srnq"] Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.160322 4770 scope.go:117] "RemoveContainer" containerID="0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.203221 4770 scope.go:117] "RemoveContainer" containerID="8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324" Oct 04 04:04:50 crc kubenswrapper[4770]: E1004 04:04:50.203884 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324\": container with ID starting with 8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324 not found: ID does not exist" containerID="8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.203945 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324"} err="failed to get container status \"8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324\": rpc error: code = NotFound desc = could not find container \"8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324\": container with ID starting with 8ea366cabf365ec1223153d2b8825ac84d0a848223d9be96bb3bf257781bc324 not found: ID does not exist" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.203984 4770 scope.go:117] "RemoveContainer" containerID="dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706" Oct 04 04:04:50 crc kubenswrapper[4770]: E1004 04:04:50.204399 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706\": container with ID starting with dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706 not found: ID does not exist" containerID="dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.204445 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706"} err="failed to get container status \"dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706\": rpc error: code = NotFound desc = could not find container \"dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706\": container with ID starting with dc7692c46e34a2d3a27297745e46a0597e5a1b97905cf13f035b5b88f520d706 not found: ID does not exist" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.204480 4770 scope.go:117] "RemoveContainer" containerID="0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62" Oct 04 04:04:50 crc kubenswrapper[4770]: E1004 04:04:50.204959 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62\": container with ID starting with 0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62 not found: ID does not exist" containerID="0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62" Oct 04 04:04:50 crc kubenswrapper[4770]: I1004 04:04:50.205049 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62"} err="failed to get container status \"0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62\": rpc error: code = NotFound desc = could not find container \"0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62\": container with ID starting with 0c662ede98a43df23bd40827c65dc065efc6f294f8e94cdeede54b4e6a87ee62 not found: ID does not exist" Oct 04 04:04:51 crc kubenswrapper[4770]: I1004 04:04:51.688181 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b50ecf50-3ddc-416e-9838-06d036a68be5" path="/var/lib/kubelet/pods/b50ecf50-3ddc-416e-9838-06d036a68be5/volumes" Oct 04 04:06:31 crc kubenswrapper[4770]: I1004 04:06:31.795402 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:06:31 crc kubenswrapper[4770]: I1004 04:06:31.796057 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:07:01 crc kubenswrapper[4770]: I1004 04:07:01.795392 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:07:01 crc kubenswrapper[4770]: I1004 04:07:01.796129 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:07:31 crc kubenswrapper[4770]: I1004 04:07:31.795295 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:07:31 crc kubenswrapper[4770]: I1004 04:07:31.795942 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:07:31 crc kubenswrapper[4770]: I1004 04:07:31.796052 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:07:31 crc kubenswrapper[4770]: I1004 04:07:31.796909 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:07:31 crc kubenswrapper[4770]: I1004 04:07:31.797035 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" gracePeriod=600 Oct 04 04:07:31 crc kubenswrapper[4770]: E1004 04:07:31.933220 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:07:32 crc kubenswrapper[4770]: I1004 04:07:32.579887 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" exitCode=0 Oct 04 04:07:32 crc kubenswrapper[4770]: I1004 04:07:32.579939 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258"} Oct 04 04:07:32 crc kubenswrapper[4770]: I1004 04:07:32.579977 4770 scope.go:117] "RemoveContainer" containerID="ef02c7e0d8b738ce4aa4b7ac3b28fc6a0bf21d37757a1d3f1e1ec6426badd23a" Oct 04 04:07:32 crc kubenswrapper[4770]: I1004 04:07:32.580527 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:07:32 crc kubenswrapper[4770]: E1004 04:07:32.580874 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:07:43 crc kubenswrapper[4770]: I1004 04:07:43.674230 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:07:43 crc kubenswrapper[4770]: E1004 04:07:43.674945 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.806442 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ntvtg"] Oct 04 04:07:55 crc kubenswrapper[4770]: E1004 04:07:55.807754 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerName="extract-utilities" Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.807789 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerName="extract-utilities" Oct 04 04:07:55 crc kubenswrapper[4770]: E1004 04:07:55.807864 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerName="registry-server" Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.807883 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerName="registry-server" Oct 04 04:07:55 crc kubenswrapper[4770]: E1004 04:07:55.807926 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerName="extract-content" Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.807946 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerName="extract-content" Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.808366 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b50ecf50-3ddc-416e-9838-06d036a68be5" containerName="registry-server" Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.810878 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.831439 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntvtg"] Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.920492 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-catalog-content\") pod \"community-operators-ntvtg\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.920889 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-utilities\") pod \"community-operators-ntvtg\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:55 crc kubenswrapper[4770]: I1004 04:07:55.920932 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94ps6\" (UniqueName: \"kubernetes.io/projected/7b548bef-1559-41ee-9b18-e2b5032fd4e4-kube-api-access-94ps6\") pod \"community-operators-ntvtg\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.021730 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94ps6\" (UniqueName: \"kubernetes.io/projected/7b548bef-1559-41ee-9b18-e2b5032fd4e4-kube-api-access-94ps6\") pod \"community-operators-ntvtg\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.021840 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-catalog-content\") pod \"community-operators-ntvtg\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.021915 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-utilities\") pod \"community-operators-ntvtg\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.022531 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-utilities\") pod \"community-operators-ntvtg\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.022568 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-catalog-content\") pod \"community-operators-ntvtg\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.049543 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94ps6\" (UniqueName: \"kubernetes.io/projected/7b548bef-1559-41ee-9b18-e2b5032fd4e4-kube-api-access-94ps6\") pod \"community-operators-ntvtg\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.156063 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.455211 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntvtg"] Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.801116 4770 generic.go:334] "Generic (PLEG): container finished" podID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerID="863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f" exitCode=0 Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.801159 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntvtg" event={"ID":"7b548bef-1559-41ee-9b18-e2b5032fd4e4","Type":"ContainerDied","Data":"863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f"} Oct 04 04:07:56 crc kubenswrapper[4770]: I1004 04:07:56.801203 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntvtg" event={"ID":"7b548bef-1559-41ee-9b18-e2b5032fd4e4","Type":"ContainerStarted","Data":"2b1b8a09cc5fc295bdfb0ee5208bff30d9cb751cb98b48117609e58c6775bb4c"} Oct 04 04:07:57 crc kubenswrapper[4770]: I1004 04:07:57.674443 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:07:57 crc kubenswrapper[4770]: E1004 04:07:57.675275 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:07:57 crc kubenswrapper[4770]: I1004 04:07:57.812337 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntvtg" event={"ID":"7b548bef-1559-41ee-9b18-e2b5032fd4e4","Type":"ContainerStarted","Data":"edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca"} Oct 04 04:07:58 crc kubenswrapper[4770]: I1004 04:07:58.820199 4770 generic.go:334] "Generic (PLEG): container finished" podID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerID="edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca" exitCode=0 Oct 04 04:07:58 crc kubenswrapper[4770]: I1004 04:07:58.820263 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntvtg" event={"ID":"7b548bef-1559-41ee-9b18-e2b5032fd4e4","Type":"ContainerDied","Data":"edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca"} Oct 04 04:07:59 crc kubenswrapper[4770]: I1004 04:07:59.832120 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntvtg" event={"ID":"7b548bef-1559-41ee-9b18-e2b5032fd4e4","Type":"ContainerStarted","Data":"bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511"} Oct 04 04:07:59 crc kubenswrapper[4770]: I1004 04:07:59.853298 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ntvtg" podStartSLOduration=2.370766508 podStartE2EDuration="4.853277307s" podCreationTimestamp="2025-10-04 04:07:55 +0000 UTC" firstStartedPulling="2025-10-04 04:07:56.80314887 +0000 UTC m=+3888.095158582" lastFinishedPulling="2025-10-04 04:07:59.285659659 +0000 UTC m=+3890.577669381" observedRunningTime="2025-10-04 04:07:59.852757974 +0000 UTC m=+3891.144767696" watchObservedRunningTime="2025-10-04 04:07:59.853277307 +0000 UTC m=+3891.145287019" Oct 04 04:08:06 crc kubenswrapper[4770]: I1004 04:08:06.156879 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:08:06 crc kubenswrapper[4770]: I1004 04:08:06.157524 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:08:06 crc kubenswrapper[4770]: I1004 04:08:06.236133 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:08:06 crc kubenswrapper[4770]: I1004 04:08:06.953872 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:08:07 crc kubenswrapper[4770]: I1004 04:08:07.009993 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntvtg"] Oct 04 04:08:08 crc kubenswrapper[4770]: I1004 04:08:08.908150 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ntvtg" podUID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerName="registry-server" containerID="cri-o://bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511" gracePeriod=2 Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.401546 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.534483 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94ps6\" (UniqueName: \"kubernetes.io/projected/7b548bef-1559-41ee-9b18-e2b5032fd4e4-kube-api-access-94ps6\") pod \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.534604 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-utilities\") pod \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.534737 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-catalog-content\") pod \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\" (UID: \"7b548bef-1559-41ee-9b18-e2b5032fd4e4\") " Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.536321 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-utilities" (OuterVolumeSpecName: "utilities") pod "7b548bef-1559-41ee-9b18-e2b5032fd4e4" (UID: "7b548bef-1559-41ee-9b18-e2b5032fd4e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.543074 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b548bef-1559-41ee-9b18-e2b5032fd4e4-kube-api-access-94ps6" (OuterVolumeSpecName: "kube-api-access-94ps6") pod "7b548bef-1559-41ee-9b18-e2b5032fd4e4" (UID: "7b548bef-1559-41ee-9b18-e2b5032fd4e4"). InnerVolumeSpecName "kube-api-access-94ps6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.619245 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b548bef-1559-41ee-9b18-e2b5032fd4e4" (UID: "7b548bef-1559-41ee-9b18-e2b5032fd4e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.637085 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94ps6\" (UniqueName: \"kubernetes.io/projected/7b548bef-1559-41ee-9b18-e2b5032fd4e4-kube-api-access-94ps6\") on node \"crc\" DevicePath \"\"" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.637145 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.637175 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b548bef-1559-41ee-9b18-e2b5032fd4e4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.920890 4770 generic.go:334] "Generic (PLEG): container finished" podID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerID="bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511" exitCode=0 Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.920937 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntvtg" event={"ID":"7b548bef-1559-41ee-9b18-e2b5032fd4e4","Type":"ContainerDied","Data":"bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511"} Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.920953 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntvtg" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.920974 4770 scope.go:117] "RemoveContainer" containerID="bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.920962 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntvtg" event={"ID":"7b548bef-1559-41ee-9b18-e2b5032fd4e4","Type":"ContainerDied","Data":"2b1b8a09cc5fc295bdfb0ee5208bff30d9cb751cb98b48117609e58c6775bb4c"} Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.963605 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntvtg"] Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.968578 4770 scope.go:117] "RemoveContainer" containerID="edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca" Oct 04 04:08:09 crc kubenswrapper[4770]: I1004 04:08:09.971314 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ntvtg"] Oct 04 04:08:10 crc kubenswrapper[4770]: I1004 04:08:10.008522 4770 scope.go:117] "RemoveContainer" containerID="863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f" Oct 04 04:08:10 crc kubenswrapper[4770]: I1004 04:08:10.027273 4770 scope.go:117] "RemoveContainer" containerID="bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511" Oct 04 04:08:10 crc kubenswrapper[4770]: E1004 04:08:10.027780 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511\": container with ID starting with bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511 not found: ID does not exist" containerID="bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511" Oct 04 04:08:10 crc kubenswrapper[4770]: I1004 04:08:10.027845 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511"} err="failed to get container status \"bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511\": rpc error: code = NotFound desc = could not find container \"bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511\": container with ID starting with bf3b23c6bf9a42fb40b23d80707a62065aa782711599c18c81e1416a87afc511 not found: ID does not exist" Oct 04 04:08:10 crc kubenswrapper[4770]: I1004 04:08:10.027936 4770 scope.go:117] "RemoveContainer" containerID="edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca" Oct 04 04:08:10 crc kubenswrapper[4770]: E1004 04:08:10.028559 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca\": container with ID starting with edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca not found: ID does not exist" containerID="edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca" Oct 04 04:08:10 crc kubenswrapper[4770]: I1004 04:08:10.028591 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca"} err="failed to get container status \"edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca\": rpc error: code = NotFound desc = could not find container \"edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca\": container with ID starting with edaaf317ea5496dedef204af431a105be1d07338c2a30daf2021ac8d761b94ca not found: ID does not exist" Oct 04 04:08:10 crc kubenswrapper[4770]: I1004 04:08:10.028617 4770 scope.go:117] "RemoveContainer" containerID="863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f" Oct 04 04:08:10 crc kubenswrapper[4770]: E1004 04:08:10.029044 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f\": container with ID starting with 863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f not found: ID does not exist" containerID="863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f" Oct 04 04:08:10 crc kubenswrapper[4770]: I1004 04:08:10.029071 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f"} err="failed to get container status \"863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f\": rpc error: code = NotFound desc = could not find container \"863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f\": container with ID starting with 863863f4b00173f2cbdf22eccf60c87461bcea16f43816621357515d881e8f4f not found: ID does not exist" Oct 04 04:08:11 crc kubenswrapper[4770]: I1004 04:08:11.689414 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" path="/var/lib/kubelet/pods/7b548bef-1559-41ee-9b18-e2b5032fd4e4/volumes" Oct 04 04:08:12 crc kubenswrapper[4770]: E1004 04:08:12.044583 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice/crio-2b1b8a09cc5fc295bdfb0ee5208bff30d9cb751cb98b48117609e58c6775bb4c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:08:12 crc kubenswrapper[4770]: I1004 04:08:12.673772 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:08:12 crc kubenswrapper[4770]: E1004 04:08:12.674589 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:08:22 crc kubenswrapper[4770]: E1004 04:08:22.267812 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice/crio-2b1b8a09cc5fc295bdfb0ee5208bff30d9cb751cb98b48117609e58c6775bb4c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:08:24 crc kubenswrapper[4770]: I1004 04:08:24.673785 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:08:24 crc kubenswrapper[4770]: E1004 04:08:24.674549 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:08:32 crc kubenswrapper[4770]: E1004 04:08:32.487148 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice/crio-2b1b8a09cc5fc295bdfb0ee5208bff30d9cb751cb98b48117609e58c6775bb4c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:08:39 crc kubenswrapper[4770]: I1004 04:08:39.683127 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:08:39 crc kubenswrapper[4770]: E1004 04:08:39.684207 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:08:42 crc kubenswrapper[4770]: E1004 04:08:42.761261 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice/crio-2b1b8a09cc5fc295bdfb0ee5208bff30d9cb751cb98b48117609e58c6775bb4c\": RecentStats: unable to find data in memory cache]" Oct 04 04:08:51 crc kubenswrapper[4770]: I1004 04:08:51.674201 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:08:51 crc kubenswrapper[4770]: E1004 04:08:51.675460 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:08:53 crc kubenswrapper[4770]: E1004 04:08:53.037823 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice/crio-2b1b8a09cc5fc295bdfb0ee5208bff30d9cb751cb98b48117609e58c6775bb4c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:09:02 crc kubenswrapper[4770]: I1004 04:09:02.673883 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:09:02 crc kubenswrapper[4770]: E1004 04:09:02.674667 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:09:03 crc kubenswrapper[4770]: E1004 04:09:03.204909 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b548bef_1559_41ee_9b18_e2b5032fd4e4.slice/crio-2b1b8a09cc5fc295bdfb0ee5208bff30d9cb751cb98b48117609e58c6775bb4c\": RecentStats: unable to find data in memory cache]" Oct 04 04:09:09 crc kubenswrapper[4770]: E1004 04:09:09.718552 4770 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/8db5d5db87b2c563c5cf6fc5d1f46ec40c3aa6731e060a9f395b2812ba7815b0/diff" to get inode usage: stat /var/lib/containers/storage/overlay/8db5d5db87b2c563c5cf6fc5d1f46ec40c3aa6731e060a9f395b2812ba7815b0/diff: no such file or directory, extraDiskErr: Oct 04 04:09:13 crc kubenswrapper[4770]: I1004 04:09:13.674394 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:09:13 crc kubenswrapper[4770]: E1004 04:09:13.675462 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:09:24 crc kubenswrapper[4770]: I1004 04:09:24.674338 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:09:24 crc kubenswrapper[4770]: E1004 04:09:24.677317 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:09:39 crc kubenswrapper[4770]: I1004 04:09:39.684167 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:09:39 crc kubenswrapper[4770]: E1004 04:09:39.685482 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:09:50 crc kubenswrapper[4770]: I1004 04:09:50.674238 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:09:50 crc kubenswrapper[4770]: E1004 04:09:50.675286 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:10:03 crc kubenswrapper[4770]: I1004 04:10:03.673776 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:10:03 crc kubenswrapper[4770]: E1004 04:10:03.674612 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:10:03 crc kubenswrapper[4770]: I1004 04:10:03.954703 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qzktq"] Oct 04 04:10:03 crc kubenswrapper[4770]: E1004 04:10:03.956989 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerName="extract-content" Oct 04 04:10:03 crc kubenswrapper[4770]: I1004 04:10:03.957060 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerName="extract-content" Oct 04 04:10:03 crc kubenswrapper[4770]: E1004 04:10:03.957102 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerName="extract-utilities" Oct 04 04:10:03 crc kubenswrapper[4770]: I1004 04:10:03.957116 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerName="extract-utilities" Oct 04 04:10:03 crc kubenswrapper[4770]: E1004 04:10:03.957160 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerName="registry-server" Oct 04 04:10:03 crc kubenswrapper[4770]: I1004 04:10:03.957173 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerName="registry-server" Oct 04 04:10:03 crc kubenswrapper[4770]: I1004 04:10:03.958477 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b548bef-1559-41ee-9b18-e2b5032fd4e4" containerName="registry-server" Oct 04 04:10:03 crc kubenswrapper[4770]: I1004 04:10:03.969806 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:03 crc kubenswrapper[4770]: I1004 04:10:03.999843 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qzktq"] Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.051036 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-969gj\" (UniqueName: \"kubernetes.io/projected/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-kube-api-access-969gj\") pod \"certified-operators-qzktq\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.051107 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-utilities\") pod \"certified-operators-qzktq\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.051205 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-catalog-content\") pod \"certified-operators-qzktq\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.152869 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-969gj\" (UniqueName: \"kubernetes.io/projected/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-kube-api-access-969gj\") pod \"certified-operators-qzktq\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.152961 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-utilities\") pod \"certified-operators-qzktq\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.153072 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-catalog-content\") pod \"certified-operators-qzktq\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.153789 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-catalog-content\") pod \"certified-operators-qzktq\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.154151 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-utilities\") pod \"certified-operators-qzktq\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.175764 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-969gj\" (UniqueName: \"kubernetes.io/projected/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-kube-api-access-969gj\") pod \"certified-operators-qzktq\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.305647 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:04 crc kubenswrapper[4770]: I1004 04:10:04.677002 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qzktq"] Oct 04 04:10:05 crc kubenswrapper[4770]: I1004 04:10:05.006464 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzktq" event={"ID":"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda","Type":"ContainerStarted","Data":"40eebb319ee05345d18ce4eb0c9a61541aa1b8ad564a80dbea6996a1cb7c6919"} Oct 04 04:10:06 crc kubenswrapper[4770]: I1004 04:10:06.031301 4770 generic.go:334] "Generic (PLEG): container finished" podID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerID="8ab53ad8478b21ca1f1ddd2bb611461d71038433ca29f0083fe1ced7b1c72ba0" exitCode=0 Oct 04 04:10:06 crc kubenswrapper[4770]: I1004 04:10:06.031362 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzktq" event={"ID":"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda","Type":"ContainerDied","Data":"8ab53ad8478b21ca1f1ddd2bb611461d71038433ca29f0083fe1ced7b1c72ba0"} Oct 04 04:10:06 crc kubenswrapper[4770]: I1004 04:10:06.034790 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:10:08 crc kubenswrapper[4770]: I1004 04:10:08.050240 4770 generic.go:334] "Generic (PLEG): container finished" podID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerID="c4c2d10b23102965c846f26671bbd328fdc85b4eda286ed1a45fbe8c1ce9c711" exitCode=0 Oct 04 04:10:08 crc kubenswrapper[4770]: I1004 04:10:08.050337 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzktq" event={"ID":"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda","Type":"ContainerDied","Data":"c4c2d10b23102965c846f26671bbd328fdc85b4eda286ed1a45fbe8c1ce9c711"} Oct 04 04:10:10 crc kubenswrapper[4770]: I1004 04:10:10.071208 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzktq" event={"ID":"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda","Type":"ContainerStarted","Data":"e57114a2f87495eac6dd8ee728b92b49244746becf45e1fbccfc0a84dfa42be8"} Oct 04 04:10:10 crc kubenswrapper[4770]: I1004 04:10:10.091924 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qzktq" podStartSLOduration=3.862593333 podStartE2EDuration="7.091908791s" podCreationTimestamp="2025-10-04 04:10:03 +0000 UTC" firstStartedPulling="2025-10-04 04:10:06.034434475 +0000 UTC m=+4017.326444197" lastFinishedPulling="2025-10-04 04:10:09.263749943 +0000 UTC m=+4020.555759655" observedRunningTime="2025-10-04 04:10:10.088530203 +0000 UTC m=+4021.380539925" watchObservedRunningTime="2025-10-04 04:10:10.091908791 +0000 UTC m=+4021.383918503" Oct 04 04:10:14 crc kubenswrapper[4770]: I1004 04:10:14.306232 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:14 crc kubenswrapper[4770]: I1004 04:10:14.308304 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:14 crc kubenswrapper[4770]: I1004 04:10:14.370604 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:15 crc kubenswrapper[4770]: I1004 04:10:15.197192 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:15 crc kubenswrapper[4770]: I1004 04:10:15.275501 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qzktq"] Oct 04 04:10:16 crc kubenswrapper[4770]: I1004 04:10:16.674948 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:10:16 crc kubenswrapper[4770]: E1004 04:10:16.675374 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:10:17 crc kubenswrapper[4770]: I1004 04:10:17.145182 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qzktq" podUID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerName="registry-server" containerID="cri-o://e57114a2f87495eac6dd8ee728b92b49244746becf45e1fbccfc0a84dfa42be8" gracePeriod=2 Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.158393 4770 generic.go:334] "Generic (PLEG): container finished" podID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerID="e57114a2f87495eac6dd8ee728b92b49244746becf45e1fbccfc0a84dfa42be8" exitCode=0 Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.158806 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzktq" event={"ID":"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda","Type":"ContainerDied","Data":"e57114a2f87495eac6dd8ee728b92b49244746becf45e1fbccfc0a84dfa42be8"} Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.259312 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.391934 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-catalog-content\") pod \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.392195 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-utilities\") pod \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.392268 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-969gj\" (UniqueName: \"kubernetes.io/projected/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-kube-api-access-969gj\") pod \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\" (UID: \"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda\") " Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.399178 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-utilities" (OuterVolumeSpecName: "utilities") pod "0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" (UID: "0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.403303 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-kube-api-access-969gj" (OuterVolumeSpecName: "kube-api-access-969gj") pod "0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" (UID: "0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda"). InnerVolumeSpecName "kube-api-access-969gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.458871 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" (UID: "0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.494553 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.494595 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-969gj\" (UniqueName: \"kubernetes.io/projected/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-kube-api-access-969gj\") on node \"crc\" DevicePath \"\"" Oct 04 04:10:18 crc kubenswrapper[4770]: I1004 04:10:18.494611 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:10:19 crc kubenswrapper[4770]: I1004 04:10:19.180076 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzktq" event={"ID":"0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda","Type":"ContainerDied","Data":"40eebb319ee05345d18ce4eb0c9a61541aa1b8ad564a80dbea6996a1cb7c6919"} Oct 04 04:10:19 crc kubenswrapper[4770]: I1004 04:10:19.180663 4770 scope.go:117] "RemoveContainer" containerID="e57114a2f87495eac6dd8ee728b92b49244746becf45e1fbccfc0a84dfa42be8" Oct 04 04:10:19 crc kubenswrapper[4770]: I1004 04:10:19.180177 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzktq" Oct 04 04:10:19 crc kubenswrapper[4770]: I1004 04:10:19.220913 4770 scope.go:117] "RemoveContainer" containerID="c4c2d10b23102965c846f26671bbd328fdc85b4eda286ed1a45fbe8c1ce9c711" Oct 04 04:10:19 crc kubenswrapper[4770]: I1004 04:10:19.241549 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qzktq"] Oct 04 04:10:19 crc kubenswrapper[4770]: I1004 04:10:19.248831 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qzktq"] Oct 04 04:10:19 crc kubenswrapper[4770]: I1004 04:10:19.253326 4770 scope.go:117] "RemoveContainer" containerID="8ab53ad8478b21ca1f1ddd2bb611461d71038433ca29f0083fe1ced7b1c72ba0" Oct 04 04:10:19 crc kubenswrapper[4770]: I1004 04:10:19.689580 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" path="/var/lib/kubelet/pods/0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda/volumes" Oct 04 04:10:29 crc kubenswrapper[4770]: I1004 04:10:29.682115 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:10:29 crc kubenswrapper[4770]: E1004 04:10:29.683224 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.634084 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tgrc5"] Oct 04 04:10:39 crc kubenswrapper[4770]: E1004 04:10:39.635203 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerName="extract-content" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.635224 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerName="extract-content" Oct 04 04:10:39 crc kubenswrapper[4770]: E1004 04:10:39.635247 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerName="registry-server" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.635258 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerName="registry-server" Oct 04 04:10:39 crc kubenswrapper[4770]: E1004 04:10:39.635284 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerName="extract-utilities" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.635295 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerName="extract-utilities" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.635525 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0204b4a6-1e1b-43c3-a3e5-9c198bc1dcda" containerName="registry-server" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.637201 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.645729 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tgrc5"] Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.777309 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-catalog-content\") pod \"redhat-operators-tgrc5\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.777473 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-utilities\") pod \"redhat-operators-tgrc5\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.777657 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcjqg\" (UniqueName: \"kubernetes.io/projected/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-kube-api-access-vcjqg\") pod \"redhat-operators-tgrc5\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.879305 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcjqg\" (UniqueName: \"kubernetes.io/projected/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-kube-api-access-vcjqg\") pod \"redhat-operators-tgrc5\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.879410 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-catalog-content\") pod \"redhat-operators-tgrc5\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.879455 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-utilities\") pod \"redhat-operators-tgrc5\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.879960 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-utilities\") pod \"redhat-operators-tgrc5\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.880155 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-catalog-content\") pod \"redhat-operators-tgrc5\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.911132 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcjqg\" (UniqueName: \"kubernetes.io/projected/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-kube-api-access-vcjqg\") pod \"redhat-operators-tgrc5\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:39 crc kubenswrapper[4770]: I1004 04:10:39.971675 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:40 crc kubenswrapper[4770]: I1004 04:10:40.417558 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tgrc5"] Oct 04 04:10:41 crc kubenswrapper[4770]: I1004 04:10:41.399799 4770 generic.go:334] "Generic (PLEG): container finished" podID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerID="c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e" exitCode=0 Oct 04 04:10:41 crc kubenswrapper[4770]: I1004 04:10:41.400232 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgrc5" event={"ID":"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63","Type":"ContainerDied","Data":"c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e"} Oct 04 04:10:41 crc kubenswrapper[4770]: I1004 04:10:41.400354 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgrc5" event={"ID":"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63","Type":"ContainerStarted","Data":"e2c9226a22a35cfa08b4f77213eae42091b5ad6fe73d1060c954ad32d639f0f7"} Oct 04 04:10:41 crc kubenswrapper[4770]: I1004 04:10:41.673553 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:10:41 crc kubenswrapper[4770]: E1004 04:10:41.673850 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:10:43 crc kubenswrapper[4770]: I1004 04:10:43.419417 4770 generic.go:334] "Generic (PLEG): container finished" podID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerID="374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf" exitCode=0 Oct 04 04:10:43 crc kubenswrapper[4770]: I1004 04:10:43.419529 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgrc5" event={"ID":"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63","Type":"ContainerDied","Data":"374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf"} Oct 04 04:10:49 crc kubenswrapper[4770]: I1004 04:10:49.473501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgrc5" event={"ID":"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63","Type":"ContainerStarted","Data":"ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c"} Oct 04 04:10:49 crc kubenswrapper[4770]: I1004 04:10:49.507169 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tgrc5" podStartSLOduration=2.761356338 podStartE2EDuration="10.507154989s" podCreationTimestamp="2025-10-04 04:10:39 +0000 UTC" firstStartedPulling="2025-10-04 04:10:41.402385453 +0000 UTC m=+4052.694395165" lastFinishedPulling="2025-10-04 04:10:49.148184064 +0000 UTC m=+4060.440193816" observedRunningTime="2025-10-04 04:10:49.505606488 +0000 UTC m=+4060.797616230" watchObservedRunningTime="2025-10-04 04:10:49.507154989 +0000 UTC m=+4060.799164701" Oct 04 04:10:49 crc kubenswrapper[4770]: I1004 04:10:49.972398 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:49 crc kubenswrapper[4770]: I1004 04:10:49.972436 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:10:51 crc kubenswrapper[4770]: I1004 04:10:51.021982 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tgrc5" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerName="registry-server" probeResult="failure" output=< Oct 04 04:10:51 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 04:10:51 crc kubenswrapper[4770]: > Oct 04 04:10:56 crc kubenswrapper[4770]: I1004 04:10:56.674394 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:10:56 crc kubenswrapper[4770]: E1004 04:10:56.675172 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:11:00 crc kubenswrapper[4770]: I1004 04:11:00.046201 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:11:00 crc kubenswrapper[4770]: I1004 04:11:00.125543 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:11:00 crc kubenswrapper[4770]: I1004 04:11:00.303332 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tgrc5"] Oct 04 04:11:01 crc kubenswrapper[4770]: I1004 04:11:01.577372 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tgrc5" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerName="registry-server" containerID="cri-o://ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c" gracePeriod=2 Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.024431 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.140750 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-catalog-content\") pod \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.140821 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-utilities\") pod \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.140885 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcjqg\" (UniqueName: \"kubernetes.io/projected/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-kube-api-access-vcjqg\") pod \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\" (UID: \"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63\") " Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.142102 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-utilities" (OuterVolumeSpecName: "utilities") pod "5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" (UID: "5a39ba2b-e39b-42f6-96d6-f8f1288d2c63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.150312 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-kube-api-access-vcjqg" (OuterVolumeSpecName: "kube-api-access-vcjqg") pod "5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" (UID: "5a39ba2b-e39b-42f6-96d6-f8f1288d2c63"). InnerVolumeSpecName "kube-api-access-vcjqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.233128 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" (UID: "5a39ba2b-e39b-42f6-96d6-f8f1288d2c63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.242242 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcjqg\" (UniqueName: \"kubernetes.io/projected/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-kube-api-access-vcjqg\") on node \"crc\" DevicePath \"\"" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.242277 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.242286 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.593316 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgrc5" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.593317 4770 generic.go:334] "Generic (PLEG): container finished" podID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerID="ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c" exitCode=0 Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.593320 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgrc5" event={"ID":"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63","Type":"ContainerDied","Data":"ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c"} Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.593469 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgrc5" event={"ID":"5a39ba2b-e39b-42f6-96d6-f8f1288d2c63","Type":"ContainerDied","Data":"e2c9226a22a35cfa08b4f77213eae42091b5ad6fe73d1060c954ad32d639f0f7"} Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.593501 4770 scope.go:117] "RemoveContainer" containerID="ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.622139 4770 scope.go:117] "RemoveContainer" containerID="374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.634794 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tgrc5"] Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.657366 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tgrc5"] Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.676403 4770 scope.go:117] "RemoveContainer" containerID="c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.698575 4770 scope.go:117] "RemoveContainer" containerID="ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c" Oct 04 04:11:02 crc kubenswrapper[4770]: E1004 04:11:02.699122 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c\": container with ID starting with ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c not found: ID does not exist" containerID="ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.699172 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c"} err="failed to get container status \"ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c\": rpc error: code = NotFound desc = could not find container \"ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c\": container with ID starting with ecfd2a7619c6f87a7d039f9af00784274ec7b56672bc5aa6bb6265c61e84d49c not found: ID does not exist" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.699205 4770 scope.go:117] "RemoveContainer" containerID="374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf" Oct 04 04:11:02 crc kubenswrapper[4770]: E1004 04:11:02.699515 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf\": container with ID starting with 374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf not found: ID does not exist" containerID="374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.699552 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf"} err="failed to get container status \"374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf\": rpc error: code = NotFound desc = could not find container \"374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf\": container with ID starting with 374e44b337f8d07c4c839625e163bb1591ffbc5929cf5b8bd9f9f41e0ba0c0cf not found: ID does not exist" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.699567 4770 scope.go:117] "RemoveContainer" containerID="c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e" Oct 04 04:11:02 crc kubenswrapper[4770]: E1004 04:11:02.699943 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e\": container with ID starting with c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e not found: ID does not exist" containerID="c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e" Oct 04 04:11:02 crc kubenswrapper[4770]: I1004 04:11:02.699988 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e"} err="failed to get container status \"c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e\": rpc error: code = NotFound desc = could not find container \"c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e\": container with ID starting with c6f979d3b09eacab73488aee4e24e115c0e93f6e1e9e5fc3864faa265ce68e5e not found: ID does not exist" Oct 04 04:11:03 crc kubenswrapper[4770]: I1004 04:11:03.685165 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" path="/var/lib/kubelet/pods/5a39ba2b-e39b-42f6-96d6-f8f1288d2c63/volumes" Oct 04 04:11:08 crc kubenswrapper[4770]: I1004 04:11:08.673668 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:11:08 crc kubenswrapper[4770]: E1004 04:11:08.674327 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:11:21 crc kubenswrapper[4770]: I1004 04:11:21.674423 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:11:21 crc kubenswrapper[4770]: E1004 04:11:21.675780 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:11:34 crc kubenswrapper[4770]: I1004 04:11:34.674335 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:11:34 crc kubenswrapper[4770]: E1004 04:11:34.675103 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:11:45 crc kubenswrapper[4770]: I1004 04:11:45.673622 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:11:45 crc kubenswrapper[4770]: E1004 04:11:45.674594 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:11:59 crc kubenswrapper[4770]: I1004 04:11:59.684881 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:11:59 crc kubenswrapper[4770]: E1004 04:11:59.688237 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:12:12 crc kubenswrapper[4770]: I1004 04:12:12.674489 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:12:12 crc kubenswrapper[4770]: E1004 04:12:12.675589 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:12:27 crc kubenswrapper[4770]: I1004 04:12:27.674621 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:12:27 crc kubenswrapper[4770]: E1004 04:12:27.675592 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:12:40 crc kubenswrapper[4770]: I1004 04:12:40.674247 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:12:41 crc kubenswrapper[4770]: I1004 04:12:41.473689 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"f3207a629dd15a66b5ea31eeb34d24cd67387bbc91230963def809efb7741eb1"} Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.781333 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vbb7r"] Oct 04 04:14:44 crc kubenswrapper[4770]: E1004 04:14:44.782279 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerName="extract-content" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.782295 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerName="extract-content" Oct 04 04:14:44 crc kubenswrapper[4770]: E1004 04:14:44.782331 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerName="registry-server" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.782340 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerName="registry-server" Oct 04 04:14:44 crc kubenswrapper[4770]: E1004 04:14:44.782360 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerName="extract-utilities" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.782369 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerName="extract-utilities" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.782537 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a39ba2b-e39b-42f6-96d6-f8f1288d2c63" containerName="registry-server" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.783866 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.810077 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbb7r"] Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.863121 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf48h\" (UniqueName: \"kubernetes.io/projected/c258ee95-2cac-4027-b5b0-cbd88947857c-kube-api-access-qf48h\") pod \"redhat-marketplace-vbb7r\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.863217 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-catalog-content\") pod \"redhat-marketplace-vbb7r\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.863296 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-utilities\") pod \"redhat-marketplace-vbb7r\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.964095 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-catalog-content\") pod \"redhat-marketplace-vbb7r\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.964186 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-utilities\") pod \"redhat-marketplace-vbb7r\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.964264 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf48h\" (UniqueName: \"kubernetes.io/projected/c258ee95-2cac-4027-b5b0-cbd88947857c-kube-api-access-qf48h\") pod \"redhat-marketplace-vbb7r\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.964923 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-catalog-content\") pod \"redhat-marketplace-vbb7r\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.965064 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-utilities\") pod \"redhat-marketplace-vbb7r\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:44 crc kubenswrapper[4770]: I1004 04:14:44.992029 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf48h\" (UniqueName: \"kubernetes.io/projected/c258ee95-2cac-4027-b5b0-cbd88947857c-kube-api-access-qf48h\") pod \"redhat-marketplace-vbb7r\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:45 crc kubenswrapper[4770]: I1004 04:14:45.106038 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:45 crc kubenswrapper[4770]: I1004 04:14:45.337055 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbb7r"] Oct 04 04:14:45 crc kubenswrapper[4770]: I1004 04:14:45.677484 4770 generic.go:334] "Generic (PLEG): container finished" podID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerID="251a11f86c0082eb75c24b6ec96e4d94451362f2e7c734c7673aded7041456cc" exitCode=0 Oct 04 04:14:45 crc kubenswrapper[4770]: I1004 04:14:45.692617 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbb7r" event={"ID":"c258ee95-2cac-4027-b5b0-cbd88947857c","Type":"ContainerDied","Data":"251a11f86c0082eb75c24b6ec96e4d94451362f2e7c734c7673aded7041456cc"} Oct 04 04:14:45 crc kubenswrapper[4770]: I1004 04:14:45.692661 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbb7r" event={"ID":"c258ee95-2cac-4027-b5b0-cbd88947857c","Type":"ContainerStarted","Data":"7e065f5877c39355c2d291358c03ff1edaf379a05d18c07d4eb25997a7e18676"} Oct 04 04:14:46 crc kubenswrapper[4770]: I1004 04:14:46.688533 4770 generic.go:334] "Generic (PLEG): container finished" podID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerID="c0e6e0daed7594b274c32a0e1accc7947b27d7531ccb370a665edbd4c4cfcad9" exitCode=0 Oct 04 04:14:46 crc kubenswrapper[4770]: I1004 04:14:46.688598 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbb7r" event={"ID":"c258ee95-2cac-4027-b5b0-cbd88947857c","Type":"ContainerDied","Data":"c0e6e0daed7594b274c32a0e1accc7947b27d7531ccb370a665edbd4c4cfcad9"} Oct 04 04:14:47 crc kubenswrapper[4770]: I1004 04:14:47.696353 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbb7r" event={"ID":"c258ee95-2cac-4027-b5b0-cbd88947857c","Type":"ContainerStarted","Data":"3edbbfef4c7a4d5764a7bc1448de8428a060ea348d12236e0a78e4685b3713a9"} Oct 04 04:14:47 crc kubenswrapper[4770]: I1004 04:14:47.718390 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vbb7r" podStartSLOduration=2.284424253 podStartE2EDuration="3.718368342s" podCreationTimestamp="2025-10-04 04:14:44 +0000 UTC" firstStartedPulling="2025-10-04 04:14:45.679521446 +0000 UTC m=+4296.971531158" lastFinishedPulling="2025-10-04 04:14:47.113465525 +0000 UTC m=+4298.405475247" observedRunningTime="2025-10-04 04:14:47.713369912 +0000 UTC m=+4299.005379624" watchObservedRunningTime="2025-10-04 04:14:47.718368342 +0000 UTC m=+4299.010378064" Oct 04 04:14:55 crc kubenswrapper[4770]: I1004 04:14:55.106211 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:55 crc kubenswrapper[4770]: I1004 04:14:55.106875 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:55 crc kubenswrapper[4770]: I1004 04:14:55.157821 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:55 crc kubenswrapper[4770]: I1004 04:14:55.831187 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:55 crc kubenswrapper[4770]: I1004 04:14:55.891152 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbb7r"] Oct 04 04:14:57 crc kubenswrapper[4770]: I1004 04:14:57.780701 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vbb7r" podUID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerName="registry-server" containerID="cri-o://3edbbfef4c7a4d5764a7bc1448de8428a060ea348d12236e0a78e4685b3713a9" gracePeriod=2 Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.792837 4770 generic.go:334] "Generic (PLEG): container finished" podID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerID="3edbbfef4c7a4d5764a7bc1448de8428a060ea348d12236e0a78e4685b3713a9" exitCode=0 Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.792918 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbb7r" event={"ID":"c258ee95-2cac-4027-b5b0-cbd88947857c","Type":"ContainerDied","Data":"3edbbfef4c7a4d5764a7bc1448de8428a060ea348d12236e0a78e4685b3713a9"} Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.793387 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbb7r" event={"ID":"c258ee95-2cac-4027-b5b0-cbd88947857c","Type":"ContainerDied","Data":"7e065f5877c39355c2d291358c03ff1edaf379a05d18c07d4eb25997a7e18676"} Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.793416 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e065f5877c39355c2d291358c03ff1edaf379a05d18c07d4eb25997a7e18676" Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.831737 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.900098 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-utilities\") pod \"c258ee95-2cac-4027-b5b0-cbd88947857c\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.900245 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-catalog-content\") pod \"c258ee95-2cac-4027-b5b0-cbd88947857c\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.900406 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf48h\" (UniqueName: \"kubernetes.io/projected/c258ee95-2cac-4027-b5b0-cbd88947857c-kube-api-access-qf48h\") pod \"c258ee95-2cac-4027-b5b0-cbd88947857c\" (UID: \"c258ee95-2cac-4027-b5b0-cbd88947857c\") " Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.900996 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-utilities" (OuterVolumeSpecName: "utilities") pod "c258ee95-2cac-4027-b5b0-cbd88947857c" (UID: "c258ee95-2cac-4027-b5b0-cbd88947857c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.912232 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c258ee95-2cac-4027-b5b0-cbd88947857c-kube-api-access-qf48h" (OuterVolumeSpecName: "kube-api-access-qf48h") pod "c258ee95-2cac-4027-b5b0-cbd88947857c" (UID: "c258ee95-2cac-4027-b5b0-cbd88947857c"). InnerVolumeSpecName "kube-api-access-qf48h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:14:58 crc kubenswrapper[4770]: I1004 04:14:58.919422 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c258ee95-2cac-4027-b5b0-cbd88947857c" (UID: "c258ee95-2cac-4027-b5b0-cbd88947857c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:14:59 crc kubenswrapper[4770]: I1004 04:14:59.002038 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf48h\" (UniqueName: \"kubernetes.io/projected/c258ee95-2cac-4027-b5b0-cbd88947857c-kube-api-access-qf48h\") on node \"crc\" DevicePath \"\"" Oct 04 04:14:59 crc kubenswrapper[4770]: I1004 04:14:59.002105 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:14:59 crc kubenswrapper[4770]: I1004 04:14:59.002117 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c258ee95-2cac-4027-b5b0-cbd88947857c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:14:59 crc kubenswrapper[4770]: I1004 04:14:59.800993 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbb7r" Oct 04 04:14:59 crc kubenswrapper[4770]: I1004 04:14:59.833844 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbb7r"] Oct 04 04:14:59 crc kubenswrapper[4770]: I1004 04:14:59.842498 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbb7r"] Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.155523 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57"] Oct 04 04:15:00 crc kubenswrapper[4770]: E1004 04:15:00.156637 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerName="registry-server" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.156657 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerName="registry-server" Oct 04 04:15:00 crc kubenswrapper[4770]: E1004 04:15:00.156691 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerName="extract-utilities" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.156699 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerName="extract-utilities" Oct 04 04:15:00 crc kubenswrapper[4770]: E1004 04:15:00.156724 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerName="extract-content" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.156731 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerName="extract-content" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.156899 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c258ee95-2cac-4027-b5b0-cbd88947857c" containerName="registry-server" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.157744 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.163245 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.163543 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.167617 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57"] Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.219876 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfzln\" (UniqueName: \"kubernetes.io/projected/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-kube-api-access-rfzln\") pod \"collect-profiles-29325855-zvz57\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.220176 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-config-volume\") pod \"collect-profiles-29325855-zvz57\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.220250 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-secret-volume\") pod \"collect-profiles-29325855-zvz57\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.322823 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-config-volume\") pod \"collect-profiles-29325855-zvz57\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.322918 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-secret-volume\") pod \"collect-profiles-29325855-zvz57\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.323027 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfzln\" (UniqueName: \"kubernetes.io/projected/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-kube-api-access-rfzln\") pod \"collect-profiles-29325855-zvz57\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.324509 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-config-volume\") pod \"collect-profiles-29325855-zvz57\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.329472 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-secret-volume\") pod \"collect-profiles-29325855-zvz57\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.644929 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfzln\" (UniqueName: \"kubernetes.io/projected/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-kube-api-access-rfzln\") pod \"collect-profiles-29325855-zvz57\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:00 crc kubenswrapper[4770]: I1004 04:15:00.787744 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:01 crc kubenswrapper[4770]: I1004 04:15:01.105818 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57"] Oct 04 04:15:01 crc kubenswrapper[4770]: I1004 04:15:01.683621 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c258ee95-2cac-4027-b5b0-cbd88947857c" path="/var/lib/kubelet/pods/c258ee95-2cac-4027-b5b0-cbd88947857c/volumes" Oct 04 04:15:01 crc kubenswrapper[4770]: I1004 04:15:01.795943 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:15:01 crc kubenswrapper[4770]: I1004 04:15:01.796042 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:15:01 crc kubenswrapper[4770]: I1004 04:15:01.825088 4770 generic.go:334] "Generic (PLEG): container finished" podID="0a4e6a8e-cf53-46f1-8726-f59fb2549eab" containerID="fbc38938c3c501a7c5ce02468bd533ba2afe4b8b4e8b2b44853e690fd57a2774" exitCode=0 Oct 04 04:15:01 crc kubenswrapper[4770]: I1004 04:15:01.825131 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" event={"ID":"0a4e6a8e-cf53-46f1-8726-f59fb2549eab","Type":"ContainerDied","Data":"fbc38938c3c501a7c5ce02468bd533ba2afe4b8b4e8b2b44853e690fd57a2774"} Oct 04 04:15:01 crc kubenswrapper[4770]: I1004 04:15:01.825159 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" event={"ID":"0a4e6a8e-cf53-46f1-8726-f59fb2549eab","Type":"ContainerStarted","Data":"9d26b1cb362ce0de9f57e153c6635a3a1d00d51099312bfe4b2991bc0484b76e"} Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.162143 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.263535 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-secret-volume\") pod \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.263622 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfzln\" (UniqueName: \"kubernetes.io/projected/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-kube-api-access-rfzln\") pod \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.263707 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-config-volume\") pod \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\" (UID: \"0a4e6a8e-cf53-46f1-8726-f59fb2549eab\") " Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.264675 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-config-volume" (OuterVolumeSpecName: "config-volume") pod "0a4e6a8e-cf53-46f1-8726-f59fb2549eab" (UID: "0a4e6a8e-cf53-46f1-8726-f59fb2549eab"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.269230 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0a4e6a8e-cf53-46f1-8726-f59fb2549eab" (UID: "0a4e6a8e-cf53-46f1-8726-f59fb2549eab"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.269241 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-kube-api-access-rfzln" (OuterVolumeSpecName: "kube-api-access-rfzln") pod "0a4e6a8e-cf53-46f1-8726-f59fb2549eab" (UID: "0a4e6a8e-cf53-46f1-8726-f59fb2549eab"). InnerVolumeSpecName "kube-api-access-rfzln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.366088 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.366554 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.366585 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfzln\" (UniqueName: \"kubernetes.io/projected/0a4e6a8e-cf53-46f1-8726-f59fb2549eab-kube-api-access-rfzln\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.843086 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" event={"ID":"0a4e6a8e-cf53-46f1-8726-f59fb2549eab","Type":"ContainerDied","Data":"9d26b1cb362ce0de9f57e153c6635a3a1d00d51099312bfe4b2991bc0484b76e"} Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.843142 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d26b1cb362ce0de9f57e153c6635a3a1d00d51099312bfe4b2991bc0484b76e" Oct 04 04:15:03 crc kubenswrapper[4770]: I1004 04:15:03.843172 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57" Oct 04 04:15:04 crc kubenswrapper[4770]: I1004 04:15:04.235585 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d"] Oct 04 04:15:04 crc kubenswrapper[4770]: I1004 04:15:04.241569 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-rtg9d"] Oct 04 04:15:05 crc kubenswrapper[4770]: I1004 04:15:05.692087 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46281317-dc20-4bc8-bd56-34683a6903e1" path="/var/lib/kubelet/pods/46281317-dc20-4bc8-bd56-34683a6903e1/volumes" Oct 04 04:15:23 crc kubenswrapper[4770]: I1004 04:15:23.856302 4770 scope.go:117] "RemoveContainer" containerID="8297f167d0a5a48392a32b35e345b2eb18440e6eb1577e963985eb48aa717c69" Oct 04 04:15:31 crc kubenswrapper[4770]: I1004 04:15:31.796370 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:15:31 crc kubenswrapper[4770]: I1004 04:15:31.797220 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:16:01 crc kubenswrapper[4770]: I1004 04:16:01.795908 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:16:01 crc kubenswrapper[4770]: I1004 04:16:01.796688 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:16:01 crc kubenswrapper[4770]: I1004 04:16:01.796752 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:16:01 crc kubenswrapper[4770]: I1004 04:16:01.797815 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3207a629dd15a66b5ea31eeb34d24cd67387bbc91230963def809efb7741eb1"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:16:01 crc kubenswrapper[4770]: I1004 04:16:01.797954 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://f3207a629dd15a66b5ea31eeb34d24cd67387bbc91230963def809efb7741eb1" gracePeriod=600 Oct 04 04:16:02 crc kubenswrapper[4770]: I1004 04:16:02.357250 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="f3207a629dd15a66b5ea31eeb34d24cd67387bbc91230963def809efb7741eb1" exitCode=0 Oct 04 04:16:02 crc kubenswrapper[4770]: I1004 04:16:02.357358 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"f3207a629dd15a66b5ea31eeb34d24cd67387bbc91230963def809efb7741eb1"} Oct 04 04:16:02 crc kubenswrapper[4770]: I1004 04:16:02.358168 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4"} Oct 04 04:16:02 crc kubenswrapper[4770]: I1004 04:16:02.358192 4770 scope.go:117] "RemoveContainer" containerID="259ccd517a77837339fe553fec871c1f8a72c4a211afae61fae90c6f09b3b258" Oct 04 04:17:04 crc kubenswrapper[4770]: I1004 04:17:04.979434 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-q2l8x"] Oct 04 04:17:04 crc kubenswrapper[4770]: I1004 04:17:04.992690 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-q2l8x"] Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.157833 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-nhzcr"] Oct 04 04:17:05 crc kubenswrapper[4770]: E1004 04:17:05.158350 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4e6a8e-cf53-46f1-8726-f59fb2549eab" containerName="collect-profiles" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.158392 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4e6a8e-cf53-46f1-8726-f59fb2549eab" containerName="collect-profiles" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.159244 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4e6a8e-cf53-46f1-8726-f59fb2549eab" containerName="collect-profiles" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.160309 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.164215 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.164308 4770 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-jz997" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.165121 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.165458 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.172845 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-nhzcr"] Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.282744 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfwff\" (UniqueName: \"kubernetes.io/projected/db88f8a4-4cf7-485f-ab59-542ed8cfed10-kube-api-access-hfwff\") pod \"crc-storage-crc-nhzcr\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.282885 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/db88f8a4-4cf7-485f-ab59-542ed8cfed10-crc-storage\") pod \"crc-storage-crc-nhzcr\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.283064 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/db88f8a4-4cf7-485f-ab59-542ed8cfed10-node-mnt\") pod \"crc-storage-crc-nhzcr\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.384644 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/db88f8a4-4cf7-485f-ab59-542ed8cfed10-node-mnt\") pod \"crc-storage-crc-nhzcr\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.384862 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfwff\" (UniqueName: \"kubernetes.io/projected/db88f8a4-4cf7-485f-ab59-542ed8cfed10-kube-api-access-hfwff\") pod \"crc-storage-crc-nhzcr\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.384985 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/db88f8a4-4cf7-485f-ab59-542ed8cfed10-crc-storage\") pod \"crc-storage-crc-nhzcr\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.385171 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/db88f8a4-4cf7-485f-ab59-542ed8cfed10-node-mnt\") pod \"crc-storage-crc-nhzcr\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.385996 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/db88f8a4-4cf7-485f-ab59-542ed8cfed10-crc-storage\") pod \"crc-storage-crc-nhzcr\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.424743 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfwff\" (UniqueName: \"kubernetes.io/projected/db88f8a4-4cf7-485f-ab59-542ed8cfed10-kube-api-access-hfwff\") pod \"crc-storage-crc-nhzcr\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.496626 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.692921 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63e7d20a-bdd8-424e-a45f-85a2a405fcb7" path="/var/lib/kubelet/pods/63e7d20a-bdd8-424e-a45f-85a2a405fcb7/volumes" Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.966940 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-nhzcr"] Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.976949 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:17:05 crc kubenswrapper[4770]: I1004 04:17:05.985897 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-nhzcr" event={"ID":"db88f8a4-4cf7-485f-ab59-542ed8cfed10","Type":"ContainerStarted","Data":"de805ea0bede597604fd8c6bfb3a5895c0f868f0951c3cbe9518c6ce12b65105"} Oct 04 04:17:06 crc kubenswrapper[4770]: I1004 04:17:06.996084 4770 generic.go:334] "Generic (PLEG): container finished" podID="db88f8a4-4cf7-485f-ab59-542ed8cfed10" containerID="4c9a5976fc39b490147d3bed73c1d8405c85bd5695e2dfa53b7d95d4222e996b" exitCode=0 Oct 04 04:17:06 crc kubenswrapper[4770]: I1004 04:17:06.996155 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-nhzcr" event={"ID":"db88f8a4-4cf7-485f-ab59-542ed8cfed10","Type":"ContainerDied","Data":"4c9a5976fc39b490147d3bed73c1d8405c85bd5695e2dfa53b7d95d4222e996b"} Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.372976 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.536551 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfwff\" (UniqueName: \"kubernetes.io/projected/db88f8a4-4cf7-485f-ab59-542ed8cfed10-kube-api-access-hfwff\") pod \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.536751 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/db88f8a4-4cf7-485f-ab59-542ed8cfed10-node-mnt\") pod \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.536823 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/db88f8a4-4cf7-485f-ab59-542ed8cfed10-crc-storage\") pod \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\" (UID: \"db88f8a4-4cf7-485f-ab59-542ed8cfed10\") " Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.536909 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db88f8a4-4cf7-485f-ab59-542ed8cfed10-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "db88f8a4-4cf7-485f-ab59-542ed8cfed10" (UID: "db88f8a4-4cf7-485f-ab59-542ed8cfed10"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.537342 4770 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/db88f8a4-4cf7-485f-ab59-542ed8cfed10-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.545297 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db88f8a4-4cf7-485f-ab59-542ed8cfed10-kube-api-access-hfwff" (OuterVolumeSpecName: "kube-api-access-hfwff") pod "db88f8a4-4cf7-485f-ab59-542ed8cfed10" (UID: "db88f8a4-4cf7-485f-ab59-542ed8cfed10"). InnerVolumeSpecName "kube-api-access-hfwff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.559286 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db88f8a4-4cf7-485f-ab59-542ed8cfed10-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "db88f8a4-4cf7-485f-ab59-542ed8cfed10" (UID: "db88f8a4-4cf7-485f-ab59-542ed8cfed10"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.638527 4770 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/db88f8a4-4cf7-485f-ab59-542ed8cfed10-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:08 crc kubenswrapper[4770]: I1004 04:17:08.638563 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfwff\" (UniqueName: \"kubernetes.io/projected/db88f8a4-4cf7-485f-ab59-542ed8cfed10-kube-api-access-hfwff\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:09 crc kubenswrapper[4770]: I1004 04:17:09.012761 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-nhzcr" event={"ID":"db88f8a4-4cf7-485f-ab59-542ed8cfed10","Type":"ContainerDied","Data":"de805ea0bede597604fd8c6bfb3a5895c0f868f0951c3cbe9518c6ce12b65105"} Oct 04 04:17:09 crc kubenswrapper[4770]: I1004 04:17:09.012849 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-nhzcr" Oct 04 04:17:09 crc kubenswrapper[4770]: I1004 04:17:09.012854 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de805ea0bede597604fd8c6bfb3a5895c0f868f0951c3cbe9518c6ce12b65105" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.000676 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-nhzcr"] Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.009440 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-nhzcr"] Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.166602 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-rzwz8"] Oct 04 04:17:11 crc kubenswrapper[4770]: E1004 04:17:11.167139 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db88f8a4-4cf7-485f-ab59-542ed8cfed10" containerName="storage" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.167172 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="db88f8a4-4cf7-485f-ab59-542ed8cfed10" containerName="storage" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.167442 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="db88f8a4-4cf7-485f-ab59-542ed8cfed10" containerName="storage" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.168286 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.170419 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.170468 4770 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-jz997" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.171302 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.180700 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.196170 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-rzwz8"] Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.274144 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbpgs\" (UniqueName: \"kubernetes.io/projected/5c946dc9-144f-449c-ae68-a4ad19903aa8-kube-api-access-pbpgs\") pod \"crc-storage-crc-rzwz8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.274236 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5c946dc9-144f-449c-ae68-a4ad19903aa8-node-mnt\") pod \"crc-storage-crc-rzwz8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.274314 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5c946dc9-144f-449c-ae68-a4ad19903aa8-crc-storage\") pod \"crc-storage-crc-rzwz8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.375818 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5c946dc9-144f-449c-ae68-a4ad19903aa8-crc-storage\") pod \"crc-storage-crc-rzwz8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.375959 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbpgs\" (UniqueName: \"kubernetes.io/projected/5c946dc9-144f-449c-ae68-a4ad19903aa8-kube-api-access-pbpgs\") pod \"crc-storage-crc-rzwz8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.376121 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5c946dc9-144f-449c-ae68-a4ad19903aa8-node-mnt\") pod \"crc-storage-crc-rzwz8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.376542 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5c946dc9-144f-449c-ae68-a4ad19903aa8-node-mnt\") pod \"crc-storage-crc-rzwz8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.377312 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5c946dc9-144f-449c-ae68-a4ad19903aa8-crc-storage\") pod \"crc-storage-crc-rzwz8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.402520 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbpgs\" (UniqueName: \"kubernetes.io/projected/5c946dc9-144f-449c-ae68-a4ad19903aa8-kube-api-access-pbpgs\") pod \"crc-storage-crc-rzwz8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.543697 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:11 crc kubenswrapper[4770]: I1004 04:17:11.686425 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db88f8a4-4cf7-485f-ab59-542ed8cfed10" path="/var/lib/kubelet/pods/db88f8a4-4cf7-485f-ab59-542ed8cfed10/volumes" Oct 04 04:17:12 crc kubenswrapper[4770]: I1004 04:17:12.000077 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-rzwz8"] Oct 04 04:17:12 crc kubenswrapper[4770]: I1004 04:17:12.041535 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-rzwz8" event={"ID":"5c946dc9-144f-449c-ae68-a4ad19903aa8","Type":"ContainerStarted","Data":"3f0994a85420781f52812b3c9790f4d45e409641e7650151ee9a14af04088198"} Oct 04 04:17:13 crc kubenswrapper[4770]: I1004 04:17:13.052679 4770 generic.go:334] "Generic (PLEG): container finished" podID="5c946dc9-144f-449c-ae68-a4ad19903aa8" containerID="c71dd9b60e788e565dac8ca31d187c091fb8aa997056524cb0b3b68718708156" exitCode=0 Oct 04 04:17:13 crc kubenswrapper[4770]: I1004 04:17:13.052739 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-rzwz8" event={"ID":"5c946dc9-144f-449c-ae68-a4ad19903aa8","Type":"ContainerDied","Data":"c71dd9b60e788e565dac8ca31d187c091fb8aa997056524cb0b3b68718708156"} Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.378468 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.523734 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbpgs\" (UniqueName: \"kubernetes.io/projected/5c946dc9-144f-449c-ae68-a4ad19903aa8-kube-api-access-pbpgs\") pod \"5c946dc9-144f-449c-ae68-a4ad19903aa8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.523846 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5c946dc9-144f-449c-ae68-a4ad19903aa8-crc-storage\") pod \"5c946dc9-144f-449c-ae68-a4ad19903aa8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.523972 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5c946dc9-144f-449c-ae68-a4ad19903aa8-node-mnt\") pod \"5c946dc9-144f-449c-ae68-a4ad19903aa8\" (UID: \"5c946dc9-144f-449c-ae68-a4ad19903aa8\") " Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.524113 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c946dc9-144f-449c-ae68-a4ad19903aa8-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "5c946dc9-144f-449c-ae68-a4ad19903aa8" (UID: "5c946dc9-144f-449c-ae68-a4ad19903aa8"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.524655 4770 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5c946dc9-144f-449c-ae68-a4ad19903aa8-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.530400 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c946dc9-144f-449c-ae68-a4ad19903aa8-kube-api-access-pbpgs" (OuterVolumeSpecName: "kube-api-access-pbpgs") pod "5c946dc9-144f-449c-ae68-a4ad19903aa8" (UID: "5c946dc9-144f-449c-ae68-a4ad19903aa8"). InnerVolumeSpecName "kube-api-access-pbpgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.548599 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c946dc9-144f-449c-ae68-a4ad19903aa8-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "5c946dc9-144f-449c-ae68-a4ad19903aa8" (UID: "5c946dc9-144f-449c-ae68-a4ad19903aa8"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.626492 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbpgs\" (UniqueName: \"kubernetes.io/projected/5c946dc9-144f-449c-ae68-a4ad19903aa8-kube-api-access-pbpgs\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:14 crc kubenswrapper[4770]: I1004 04:17:14.626558 4770 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5c946dc9-144f-449c-ae68-a4ad19903aa8-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:15 crc kubenswrapper[4770]: I1004 04:17:15.074140 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-rzwz8" event={"ID":"5c946dc9-144f-449c-ae68-a4ad19903aa8","Type":"ContainerDied","Data":"3f0994a85420781f52812b3c9790f4d45e409641e7650151ee9a14af04088198"} Oct 04 04:17:15 crc kubenswrapper[4770]: I1004 04:17:15.074221 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f0994a85420781f52812b3c9790f4d45e409641e7650151ee9a14af04088198" Oct 04 04:17:15 crc kubenswrapper[4770]: I1004 04:17:15.074225 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-rzwz8" Oct 04 04:17:23 crc kubenswrapper[4770]: I1004 04:17:23.950417 4770 scope.go:117] "RemoveContainer" containerID="512ccd5465fdf014c659aa03b4fce7908acc47e9ff5722ef319b6a583f9c0c78" Oct 04 04:18:31 crc kubenswrapper[4770]: I1004 04:18:31.796253 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:18:31 crc kubenswrapper[4770]: I1004 04:18:31.798703 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:19:01 crc kubenswrapper[4770]: I1004 04:19:01.796498 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:19:01 crc kubenswrapper[4770]: I1004 04:19:01.797475 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:19:31 crc kubenswrapper[4770]: I1004 04:19:31.795574 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:19:31 crc kubenswrapper[4770]: I1004 04:19:31.796415 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:19:31 crc kubenswrapper[4770]: I1004 04:19:31.796516 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:19:31 crc kubenswrapper[4770]: I1004 04:19:31.797474 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:19:31 crc kubenswrapper[4770]: I1004 04:19:31.797579 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" gracePeriod=600 Oct 04 04:19:31 crc kubenswrapper[4770]: E1004 04:19:31.946292 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:19:32 crc kubenswrapper[4770]: I1004 04:19:32.334702 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" exitCode=0 Oct 04 04:19:32 crc kubenswrapper[4770]: I1004 04:19:32.334775 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4"} Oct 04 04:19:32 crc kubenswrapper[4770]: I1004 04:19:32.335090 4770 scope.go:117] "RemoveContainer" containerID="f3207a629dd15a66b5ea31eeb34d24cd67387bbc91230963def809efb7741eb1" Oct 04 04:19:32 crc kubenswrapper[4770]: I1004 04:19:32.336240 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:19:32 crc kubenswrapper[4770]: E1004 04:19:32.336678 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:19:42 crc kubenswrapper[4770]: I1004 04:19:42.673868 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:19:42 crc kubenswrapper[4770]: E1004 04:19:42.674825 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:19:55 crc kubenswrapper[4770]: I1004 04:19:55.674789 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:19:55 crc kubenswrapper[4770]: E1004 04:19:55.676129 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:20:10 crc kubenswrapper[4770]: I1004 04:20:10.674273 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:20:10 crc kubenswrapper[4770]: E1004 04:20:10.675264 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:20:25 crc kubenswrapper[4770]: I1004 04:20:25.673400 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:20:25 crc kubenswrapper[4770]: E1004 04:20:25.674524 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.577828 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56b49fb57-9l74z"] Oct 04 04:20:31 crc kubenswrapper[4770]: E1004 04:20:31.578716 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c946dc9-144f-449c-ae68-a4ad19903aa8" containerName="storage" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.578729 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c946dc9-144f-449c-ae68-a4ad19903aa8" containerName="storage" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.578882 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c946dc9-144f-449c-ae68-a4ad19903aa8" containerName="storage" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.579791 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.582264 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rvx4c" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.582435 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.582627 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.583160 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.585494 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.588951 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56b49fb57-9l74z"] Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.747847 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzlsk\" (UniqueName: \"kubernetes.io/projected/278b89af-8ec2-4258-b7d2-d41e2e58df86-kube-api-access-zzlsk\") pod \"dnsmasq-dns-56b49fb57-9l74z\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.747943 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-config\") pod \"dnsmasq-dns-56b49fb57-9l74z\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.747969 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-dns-svc\") pod \"dnsmasq-dns-56b49fb57-9l74z\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.849047 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-config\") pod \"dnsmasq-dns-56b49fb57-9l74z\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.849099 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-dns-svc\") pod \"dnsmasq-dns-56b49fb57-9l74z\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.849179 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzlsk\" (UniqueName: \"kubernetes.io/projected/278b89af-8ec2-4258-b7d2-d41e2e58df86-kube-api-access-zzlsk\") pod \"dnsmasq-dns-56b49fb57-9l74z\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.850718 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-dns-svc\") pod \"dnsmasq-dns-56b49fb57-9l74z\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.851073 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-config\") pod \"dnsmasq-dns-56b49fb57-9l74z\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.877190 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzlsk\" (UniqueName: \"kubernetes.io/projected/278b89af-8ec2-4258-b7d2-d41e2e58df86-kube-api-access-zzlsk\") pod \"dnsmasq-dns-56b49fb57-9l74z\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.885850 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64547bf95c-lm8q9"] Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.887147 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.900346 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64547bf95c-lm8q9"] Oct 04 04:20:31 crc kubenswrapper[4770]: I1004 04:20:31.911335 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.052115 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-dns-svc\") pod \"dnsmasq-dns-64547bf95c-lm8q9\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.052469 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-config\") pod \"dnsmasq-dns-64547bf95c-lm8q9\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.052497 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg4m8\" (UniqueName: \"kubernetes.io/projected/5c978011-2d08-4128-a442-4f4e59ff7587-kube-api-access-dg4m8\") pod \"dnsmasq-dns-64547bf95c-lm8q9\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.159426 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-dns-svc\") pod \"dnsmasq-dns-64547bf95c-lm8q9\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.159479 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-config\") pod \"dnsmasq-dns-64547bf95c-lm8q9\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.159499 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg4m8\" (UniqueName: \"kubernetes.io/projected/5c978011-2d08-4128-a442-4f4e59ff7587-kube-api-access-dg4m8\") pod \"dnsmasq-dns-64547bf95c-lm8q9\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.160734 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-dns-svc\") pod \"dnsmasq-dns-64547bf95c-lm8q9\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.161062 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-config\") pod \"dnsmasq-dns-64547bf95c-lm8q9\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.187791 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg4m8\" (UniqueName: \"kubernetes.io/projected/5c978011-2d08-4128-a442-4f4e59ff7587-kube-api-access-dg4m8\") pod \"dnsmasq-dns-64547bf95c-lm8q9\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.250400 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.421623 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56b49fb57-9l74z"] Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.477020 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64547bf95c-lm8q9"] Oct 04 04:20:32 crc kubenswrapper[4770]: W1004 04:20:32.488709 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c978011_2d08_4128_a442_4f4e59ff7587.slice/crio-755178700ca8eb8ec48c5d580d7149abb2e97d2639d6f88832fb902f743bb035 WatchSource:0}: Error finding container 755178700ca8eb8ec48c5d580d7149abb2e97d2639d6f88832fb902f743bb035: Status 404 returned error can't find the container with id 755178700ca8eb8ec48c5d580d7149abb2e97d2639d6f88832fb902f743bb035 Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.751470 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.753633 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.760872 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.761112 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.760965 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.761062 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6z9fp" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.766144 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.785876 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.870488 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.870539 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.870563 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.870616 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpgx5\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-kube-api-access-vpgx5\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.870737 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.870856 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.871042 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.871120 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.871291 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.902253 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" event={"ID":"5c978011-2d08-4128-a442-4f4e59ff7587","Type":"ContainerStarted","Data":"755178700ca8eb8ec48c5d580d7149abb2e97d2639d6f88832fb902f743bb035"} Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.904700 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" event={"ID":"278b89af-8ec2-4258-b7d2-d41e2e58df86","Type":"ContainerStarted","Data":"deb19ad8d68535955132ad6f2accac4872280c4a7f66f0424ba5671d99f45f55"} Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.974080 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.974668 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.974790 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.974872 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpgx5\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-kube-api-access-vpgx5\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.974980 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.975072 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.975156 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.975221 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.975291 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.975798 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.976374 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.977158 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.977253 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.977478 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:20:32 crc kubenswrapper[4770]: I1004 04:20:32.977584 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/906571f9b5ef483062810d0a89b78f264350f028b315b995e454b2cd8a5402c8/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.046955 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.047310 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.048530 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.049610 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpgx5\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-kube-api-access-vpgx5\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.064548 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.066570 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.070348 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.070857 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.071254 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qlcz4" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.074551 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.077720 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.079242 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.095287 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") pod \"rabbitmq-server-0\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " pod="openstack/rabbitmq-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.177988 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvgbn\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-kube-api-access-hvgbn\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.178061 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.178096 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.178125 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.178142 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/661cb72b-18eb-47fc-94b3-53fa8c50aec8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.178177 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.178198 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.178218 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/661cb72b-18eb-47fc-94b3-53fa8c50aec8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.178239 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.286521 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.286578 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/661cb72b-18eb-47fc-94b3-53fa8c50aec8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.286610 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.286674 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvgbn\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-kube-api-access-hvgbn\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.286717 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.286758 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.286786 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.286809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/661cb72b-18eb-47fc-94b3-53fa8c50aec8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.286870 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.287804 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.288617 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.289836 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.294980 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.299722 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/661cb72b-18eb-47fc-94b3-53fa8c50aec8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.300210 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.307999 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/be312a493dc81d83360396620ad55cfc993563ccce411951daf7b4e05473c6be/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.301258 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/661cb72b-18eb-47fc-94b3-53fa8c50aec8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.301610 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.317436 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvgbn\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-kube-api-access-hvgbn\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.392861 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.472366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.496416 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:20:33 crc kubenswrapper[4770]: W1004 04:20:33.816411 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f38a480_df1d_4f8c_8009_60f12b7fdb2f.slice/crio-af30070c951a239d48949cdfe825688b82ab3001ee4533f340e07b6cca7cd56b WatchSource:0}: Error finding container af30070c951a239d48949cdfe825688b82ab3001ee4533f340e07b6cca7cd56b: Status 404 returned error can't find the container with id af30070c951a239d48949cdfe825688b82ab3001ee4533f340e07b6cca7cd56b Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.817833 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.916329 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8f38a480-df1d-4f8c-8009-60f12b7fdb2f","Type":"ContainerStarted","Data":"af30070c951a239d48949cdfe825688b82ab3001ee4533f340e07b6cca7cd56b"} Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.919144 4770 generic.go:334] "Generic (PLEG): container finished" podID="5c978011-2d08-4128-a442-4f4e59ff7587" containerID="8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac" exitCode=0 Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.919243 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" event={"ID":"5c978011-2d08-4128-a442-4f4e59ff7587","Type":"ContainerDied","Data":"8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac"} Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.922972 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.923041 4770 generic.go:334] "Generic (PLEG): container finished" podID="278b89af-8ec2-4258-b7d2-d41e2e58df86" containerID="12a7dc4fd40ea309fae88dd2cf064509e7fa7e99ba5efebbc2dd1d4b02afa6ed" exitCode=0 Oct 04 04:20:33 crc kubenswrapper[4770]: I1004 04:20:33.923100 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" event={"ID":"278b89af-8ec2-4258-b7d2-d41e2e58df86","Type":"ContainerDied","Data":"12a7dc4fd40ea309fae88dd2cf064509e7fa7e99ba5efebbc2dd1d4b02afa6ed"} Oct 04 04:20:33 crc kubenswrapper[4770]: W1004 04:20:33.925558 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod661cb72b_18eb_47fc_94b3_53fa8c50aec8.slice/crio-e9874c42f82487ac92586e4d2dd486408067a0d7d1da1bdc4e66dd7fdfb17d39 WatchSource:0}: Error finding container e9874c42f82487ac92586e4d2dd486408067a0d7d1da1bdc4e66dd7fdfb17d39: Status 404 returned error can't find the container with id e9874c42f82487ac92586e4d2dd486408067a0d7d1da1bdc4e66dd7fdfb17d39 Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.424418 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.426652 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.429773 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.430133 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.430296 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.430475 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-hrfbl" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.430624 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.436600 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.440070 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.507996 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wj9x\" (UniqueName: \"kubernetes.io/projected/3bceaca0-f90b-4feb-b54c-6292db9d580c-kube-api-access-5wj9x\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.508107 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bceaca0-f90b-4feb-b54c-6292db9d580c-kolla-config\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.508203 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3bceaca0-f90b-4feb-b54c-6292db9d580c-config-data-default\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.508245 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bceaca0-f90b-4feb-b54c-6292db9d580c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.508280 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a2f92edb-50e3-4cc6-911f-0d188f1771af\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2f92edb-50e3-4cc6-911f-0d188f1771af\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.508307 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bceaca0-f90b-4feb-b54c-6292db9d580c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.508336 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3bceaca0-f90b-4feb-b54c-6292db9d580c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.508384 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bceaca0-f90b-4feb-b54c-6292db9d580c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.508430 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3bceaca0-f90b-4feb-b54c-6292db9d580c-secrets\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.549542 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.550738 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.552692 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-gjlmn" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.553331 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.564588 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.610222 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bceaca0-f90b-4feb-b54c-6292db9d580c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.611034 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3bceaca0-f90b-4feb-b54c-6292db9d580c-secrets\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.611078 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wj9x\" (UniqueName: \"kubernetes.io/projected/3bceaca0-f90b-4feb-b54c-6292db9d580c-kube-api-access-5wj9x\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.611132 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bceaca0-f90b-4feb-b54c-6292db9d580c-kolla-config\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.611202 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3bceaca0-f90b-4feb-b54c-6292db9d580c-config-data-default\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.611233 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bceaca0-f90b-4feb-b54c-6292db9d580c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.611262 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a2f92edb-50e3-4cc6-911f-0d188f1771af\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2f92edb-50e3-4cc6-911f-0d188f1771af\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.611285 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bceaca0-f90b-4feb-b54c-6292db9d580c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.611314 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3bceaca0-f90b-4feb-b54c-6292db9d580c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.611763 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3bceaca0-f90b-4feb-b54c-6292db9d580c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.612887 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3bceaca0-f90b-4feb-b54c-6292db9d580c-config-data-default\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.613545 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bceaca0-f90b-4feb-b54c-6292db9d580c-kolla-config\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.613751 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bceaca0-f90b-4feb-b54c-6292db9d580c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.616213 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bceaca0-f90b-4feb-b54c-6292db9d580c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.617410 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bceaca0-f90b-4feb-b54c-6292db9d580c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.617733 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3bceaca0-f90b-4feb-b54c-6292db9d580c-secrets\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.617785 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.617809 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a2f92edb-50e3-4cc6-911f-0d188f1771af\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2f92edb-50e3-4cc6-911f-0d188f1771af\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ff978bbd17d31917b57ce2e6cdd76b1a5436a46ab7d4c614dc4af79bfc8ae72e/globalmount\"" pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.631519 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wj9x\" (UniqueName: \"kubernetes.io/projected/3bceaca0-f90b-4feb-b54c-6292db9d580c-kube-api-access-5wj9x\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.674558 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a2f92edb-50e3-4cc6-911f-0d188f1771af\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2f92edb-50e3-4cc6-911f-0d188f1771af\") pod \"openstack-galera-0\" (UID: \"3bceaca0-f90b-4feb-b54c-6292db9d580c\") " pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.713129 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcccm\" (UniqueName: \"kubernetes.io/projected/b430eca9-a710-4be3-b5ee-228951137298-kube-api-access-bcccm\") pod \"memcached-0\" (UID: \"b430eca9-a710-4be3-b5ee-228951137298\") " pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.713194 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b430eca9-a710-4be3-b5ee-228951137298-config-data\") pod \"memcached-0\" (UID: \"b430eca9-a710-4be3-b5ee-228951137298\") " pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.713227 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b430eca9-a710-4be3-b5ee-228951137298-kolla-config\") pod \"memcached-0\" (UID: \"b430eca9-a710-4be3-b5ee-228951137298\") " pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.744746 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.814816 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcccm\" (UniqueName: \"kubernetes.io/projected/b430eca9-a710-4be3-b5ee-228951137298-kube-api-access-bcccm\") pod \"memcached-0\" (UID: \"b430eca9-a710-4be3-b5ee-228951137298\") " pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.814909 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b430eca9-a710-4be3-b5ee-228951137298-config-data\") pod \"memcached-0\" (UID: \"b430eca9-a710-4be3-b5ee-228951137298\") " pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.814941 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b430eca9-a710-4be3-b5ee-228951137298-kolla-config\") pod \"memcached-0\" (UID: \"b430eca9-a710-4be3-b5ee-228951137298\") " pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.815871 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b430eca9-a710-4be3-b5ee-228951137298-config-data\") pod \"memcached-0\" (UID: \"b430eca9-a710-4be3-b5ee-228951137298\") " pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.815917 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b430eca9-a710-4be3-b5ee-228951137298-kolla-config\") pod \"memcached-0\" (UID: \"b430eca9-a710-4be3-b5ee-228951137298\") " pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.833174 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcccm\" (UniqueName: \"kubernetes.io/projected/b430eca9-a710-4be3-b5ee-228951137298-kube-api-access-bcccm\") pod \"memcached-0\" (UID: \"b430eca9-a710-4be3-b5ee-228951137298\") " pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.867999 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.933676 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" event={"ID":"5c978011-2d08-4128-a442-4f4e59ff7587","Type":"ContainerStarted","Data":"64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896"} Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.933939 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.941407 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" event={"ID":"278b89af-8ec2-4258-b7d2-d41e2e58df86","Type":"ContainerStarted","Data":"d18e8df336727677ec6aeb865e1e8ee1225312bcb37020f7c008f7cf044489c6"} Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.942066 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.943634 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"661cb72b-18eb-47fc-94b3-53fa8c50aec8","Type":"ContainerStarted","Data":"e9874c42f82487ac92586e4d2dd486408067a0d7d1da1bdc4e66dd7fdfb17d39"} Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.958738 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" podStartSLOduration=3.958720642 podStartE2EDuration="3.958720642s" podCreationTimestamp="2025-10-04 04:20:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:20:34.950887617 +0000 UTC m=+4646.242897329" watchObservedRunningTime="2025-10-04 04:20:34.958720642 +0000 UTC m=+4646.250730354" Oct 04 04:20:34 crc kubenswrapper[4770]: I1004 04:20:34.983812 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" podStartSLOduration=3.983794987 podStartE2EDuration="3.983794987s" podCreationTimestamp="2025-10-04 04:20:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:20:34.982640017 +0000 UTC m=+4646.274649729" watchObservedRunningTime="2025-10-04 04:20:34.983794987 +0000 UTC m=+4646.275804699" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.167469 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.300905 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 04:20:35 crc kubenswrapper[4770]: W1004 04:20:35.449252 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb430eca9_a710_4be3_b5ee_228951137298.slice/crio-234b1fa2e1b2434ab649fb07b174d9f1c05b8e2269d674c76a85521f3dda3ccc WatchSource:0}: Error finding container 234b1fa2e1b2434ab649fb07b174d9f1c05b8e2269d674c76a85521f3dda3ccc: Status 404 returned error can't find the container with id 234b1fa2e1b2434ab649fb07b174d9f1c05b8e2269d674c76a85521f3dda3ccc Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.746799 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.752146 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.757576 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.758776 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-pp4mb" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.759959 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.761881 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.788343 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.827131 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.827174 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.827196 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.827217 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fb7d743c-92e2-4004-878a-0ab357467b59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb7d743c-92e2-4004-878a-0ab357467b59\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.827298 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w4p2\" (UniqueName: \"kubernetes.io/projected/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-kube-api-access-7w4p2\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.827357 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.827519 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.827574 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.827648 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.929212 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.929294 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.929370 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.929413 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.929450 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.929485 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.929518 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fb7d743c-92e2-4004-878a-0ab357467b59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb7d743c-92e2-4004-878a-0ab357467b59\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.929550 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w4p2\" (UniqueName: \"kubernetes.io/projected/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-kube-api-access-7w4p2\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.929588 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.930627 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.930994 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.931250 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.931751 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.934310 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.934356 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fb7d743c-92e2-4004-878a-0ab357467b59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb7d743c-92e2-4004-878a-0ab357467b59\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/dd4133d3163a9ab7caea4014b8dc4869f07bee702a27c2e0a7f93df7ede900c4/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.935814 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.936725 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.939737 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.956571 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w4p2\" (UniqueName: \"kubernetes.io/projected/e130aa71-bdc5-49f4-8fa4-d2e96cbf359e-kube-api-access-7w4p2\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.957883 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8f38a480-df1d-4f8c-8009-60f12b7fdb2f","Type":"ContainerStarted","Data":"734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d"} Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.960194 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"661cb72b-18eb-47fc-94b3-53fa8c50aec8","Type":"ContainerStarted","Data":"6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c"} Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.963745 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b430eca9-a710-4be3-b5ee-228951137298","Type":"ContainerStarted","Data":"cffa9894a7e0e7d918ca89e81aa2663877752414339da89399e2eb5b69ae373b"} Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.967708 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b430eca9-a710-4be3-b5ee-228951137298","Type":"ContainerStarted","Data":"234b1fa2e1b2434ab649fb07b174d9f1c05b8e2269d674c76a85521f3dda3ccc"} Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.967826 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.967841 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3bceaca0-f90b-4feb-b54c-6292db9d580c","Type":"ContainerStarted","Data":"f84e3ebc142c0c1488acb089e4601e8ef2ad994f8444a7e33491253e20818b38"} Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.967853 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3bceaca0-f90b-4feb-b54c-6292db9d580c","Type":"ContainerStarted","Data":"f79f661bb4be2fde4754d9bd6be8a1ecf60bf9139f5af1aa2ffaa8d6975e438d"} Oct 04 04:20:35 crc kubenswrapper[4770]: I1004 04:20:35.976050 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fb7d743c-92e2-4004-878a-0ab357467b59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb7d743c-92e2-4004-878a-0ab357467b59\") pod \"openstack-cell1-galera-0\" (UID: \"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:36 crc kubenswrapper[4770]: I1004 04:20:36.032617 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.032593318 podStartE2EDuration="2.032593318s" podCreationTimestamp="2025-10-04 04:20:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:20:36.02579786 +0000 UTC m=+4647.317807602" watchObservedRunningTime="2025-10-04 04:20:36.032593318 +0000 UTC m=+4647.324603060" Oct 04 04:20:36 crc kubenswrapper[4770]: I1004 04:20:36.092080 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:36 crc kubenswrapper[4770]: I1004 04:20:36.602491 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 04:20:36 crc kubenswrapper[4770]: W1004 04:20:36.609608 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode130aa71_bdc5_49f4_8fa4_d2e96cbf359e.slice/crio-2e8f5d4de9a17c259e6cc246ca462f0b22970f6128c5089a1c3ba61e26d0833a WatchSource:0}: Error finding container 2e8f5d4de9a17c259e6cc246ca462f0b22970f6128c5089a1c3ba61e26d0833a: Status 404 returned error can't find the container with id 2e8f5d4de9a17c259e6cc246ca462f0b22970f6128c5089a1c3ba61e26d0833a Oct 04 04:20:36 crc kubenswrapper[4770]: I1004 04:20:36.977615 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e","Type":"ContainerStarted","Data":"9a77b0b802a56e3df27bb5716a743f624f6f31ad92b9e2d5f8b5af4cb420907f"} Oct 04 04:20:36 crc kubenswrapper[4770]: I1004 04:20:36.979397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e","Type":"ContainerStarted","Data":"2e8f5d4de9a17c259e6cc246ca462f0b22970f6128c5089a1c3ba61e26d0833a"} Oct 04 04:20:40 crc kubenswrapper[4770]: I1004 04:20:40.673519 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:20:40 crc kubenswrapper[4770]: E1004 04:20:40.675971 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:20:41 crc kubenswrapper[4770]: I1004 04:20:41.913325 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:42 crc kubenswrapper[4770]: I1004 04:20:42.252308 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:20:42 crc kubenswrapper[4770]: I1004 04:20:42.307080 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56b49fb57-9l74z"] Oct 04 04:20:42 crc kubenswrapper[4770]: I1004 04:20:42.307273 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" podUID="278b89af-8ec2-4258-b7d2-d41e2e58df86" containerName="dnsmasq-dns" containerID="cri-o://d18e8df336727677ec6aeb865e1e8ee1225312bcb37020f7c008f7cf044489c6" gracePeriod=10 Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.042715 4770 generic.go:334] "Generic (PLEG): container finished" podID="278b89af-8ec2-4258-b7d2-d41e2e58df86" containerID="d18e8df336727677ec6aeb865e1e8ee1225312bcb37020f7c008f7cf044489c6" exitCode=0 Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.044150 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" event={"ID":"278b89af-8ec2-4258-b7d2-d41e2e58df86","Type":"ContainerDied","Data":"d18e8df336727677ec6aeb865e1e8ee1225312bcb37020f7c008f7cf044489c6"} Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.343156 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.462603 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzlsk\" (UniqueName: \"kubernetes.io/projected/278b89af-8ec2-4258-b7d2-d41e2e58df86-kube-api-access-zzlsk\") pod \"278b89af-8ec2-4258-b7d2-d41e2e58df86\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.462678 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-config\") pod \"278b89af-8ec2-4258-b7d2-d41e2e58df86\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.462928 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-dns-svc\") pod \"278b89af-8ec2-4258-b7d2-d41e2e58df86\" (UID: \"278b89af-8ec2-4258-b7d2-d41e2e58df86\") " Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.474281 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/278b89af-8ec2-4258-b7d2-d41e2e58df86-kube-api-access-zzlsk" (OuterVolumeSpecName: "kube-api-access-zzlsk") pod "278b89af-8ec2-4258-b7d2-d41e2e58df86" (UID: "278b89af-8ec2-4258-b7d2-d41e2e58df86"). InnerVolumeSpecName "kube-api-access-zzlsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.498767 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-config" (OuterVolumeSpecName: "config") pod "278b89af-8ec2-4258-b7d2-d41e2e58df86" (UID: "278b89af-8ec2-4258-b7d2-d41e2e58df86"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.524059 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "278b89af-8ec2-4258-b7d2-d41e2e58df86" (UID: "278b89af-8ec2-4258-b7d2-d41e2e58df86"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.564913 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.564964 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzlsk\" (UniqueName: \"kubernetes.io/projected/278b89af-8ec2-4258-b7d2-d41e2e58df86-kube-api-access-zzlsk\") on node \"crc\" DevicePath \"\"" Oct 04 04:20:43 crc kubenswrapper[4770]: I1004 04:20:43.564978 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/278b89af-8ec2-4258-b7d2-d41e2e58df86-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:20:44 crc kubenswrapper[4770]: I1004 04:20:44.058341 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" event={"ID":"278b89af-8ec2-4258-b7d2-d41e2e58df86","Type":"ContainerDied","Data":"deb19ad8d68535955132ad6f2accac4872280c4a7f66f0424ba5671d99f45f55"} Oct 04 04:20:44 crc kubenswrapper[4770]: I1004 04:20:44.058418 4770 scope.go:117] "RemoveContainer" containerID="d18e8df336727677ec6aeb865e1e8ee1225312bcb37020f7c008f7cf044489c6" Oct 04 04:20:44 crc kubenswrapper[4770]: I1004 04:20:44.058620 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56b49fb57-9l74z" Oct 04 04:20:44 crc kubenswrapper[4770]: I1004 04:20:44.088151 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56b49fb57-9l74z"] Oct 04 04:20:44 crc kubenswrapper[4770]: I1004 04:20:44.092795 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56b49fb57-9l74z"] Oct 04 04:20:44 crc kubenswrapper[4770]: I1004 04:20:44.093124 4770 scope.go:117] "RemoveContainer" containerID="12a7dc4fd40ea309fae88dd2cf064509e7fa7e99ba5efebbc2dd1d4b02afa6ed" Oct 04 04:20:44 crc kubenswrapper[4770]: I1004 04:20:44.869966 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 04 04:20:45 crc kubenswrapper[4770]: I1004 04:20:45.689358 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="278b89af-8ec2-4258-b7d2-d41e2e58df86" path="/var/lib/kubelet/pods/278b89af-8ec2-4258-b7d2-d41e2e58df86/volumes" Oct 04 04:20:50 crc kubenswrapper[4770]: I1004 04:20:50.114273 4770 generic.go:334] "Generic (PLEG): container finished" podID="3bceaca0-f90b-4feb-b54c-6292db9d580c" containerID="f84e3ebc142c0c1488acb089e4601e8ef2ad994f8444a7e33491253e20818b38" exitCode=0 Oct 04 04:20:50 crc kubenswrapper[4770]: I1004 04:20:50.114376 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3bceaca0-f90b-4feb-b54c-6292db9d580c","Type":"ContainerDied","Data":"f84e3ebc142c0c1488acb089e4601e8ef2ad994f8444a7e33491253e20818b38"} Oct 04 04:20:50 crc kubenswrapper[4770]: I1004 04:20:50.125272 4770 generic.go:334] "Generic (PLEG): container finished" podID="e130aa71-bdc5-49f4-8fa4-d2e96cbf359e" containerID="9a77b0b802a56e3df27bb5716a743f624f6f31ad92b9e2d5f8b5af4cb420907f" exitCode=0 Oct 04 04:20:50 crc kubenswrapper[4770]: I1004 04:20:50.125340 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e","Type":"ContainerDied","Data":"9a77b0b802a56e3df27bb5716a743f624f6f31ad92b9e2d5f8b5af4cb420907f"} Oct 04 04:20:51 crc kubenswrapper[4770]: I1004 04:20:51.139270 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3bceaca0-f90b-4feb-b54c-6292db9d580c","Type":"ContainerStarted","Data":"44cdd2dde9f08769f6daedf767f634fb5056caf59025919de09293abb2187458"} Oct 04 04:20:51 crc kubenswrapper[4770]: I1004 04:20:51.144138 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e130aa71-bdc5-49f4-8fa4-d2e96cbf359e","Type":"ContainerStarted","Data":"ca7c2252ab777bbe141bab9e0fcbd512b4878cf03cced218966583d26594499e"} Oct 04 04:20:51 crc kubenswrapper[4770]: I1004 04:20:51.179727 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=18.179697603 podStartE2EDuration="18.179697603s" podCreationTimestamp="2025-10-04 04:20:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:20:51.177384932 +0000 UTC m=+4662.469394674" watchObservedRunningTime="2025-10-04 04:20:51.179697603 +0000 UTC m=+4662.471707345" Oct 04 04:20:51 crc kubenswrapper[4770]: I1004 04:20:51.212318 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=17.212282403 podStartE2EDuration="17.212282403s" podCreationTimestamp="2025-10-04 04:20:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:20:51.206720468 +0000 UTC m=+4662.498730220" watchObservedRunningTime="2025-10-04 04:20:51.212282403 +0000 UTC m=+4662.504292155" Oct 04 04:20:52 crc kubenswrapper[4770]: I1004 04:20:52.673296 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:20:52 crc kubenswrapper[4770]: E1004 04:20:52.673784 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:20:53 crc kubenswrapper[4770]: E1004 04:20:53.870268 4770 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.68:38708->38.102.83.68:36801: write tcp 38.102.83.68:38708->38.102.83.68:36801: write: broken pipe Oct 04 04:20:54 crc kubenswrapper[4770]: I1004 04:20:54.745842 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 04 04:20:54 crc kubenswrapper[4770]: I1004 04:20:54.745909 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 04 04:20:56 crc kubenswrapper[4770]: I1004 04:20:56.093399 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:56 crc kubenswrapper[4770]: I1004 04:20:56.093927 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 04 04:20:56 crc kubenswrapper[4770]: I1004 04:20:56.905862 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 04 04:20:56 crc kubenswrapper[4770]: I1004 04:20:56.991458 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="3bceaca0-f90b-4feb-b54c-6292db9d580c" containerName="galera" probeResult="failure" output=< Oct 04 04:20:56 crc kubenswrapper[4770]: wsrep_local_state_comment (Joined) differs from Synced Oct 04 04:20:56 crc kubenswrapper[4770]: > Oct 04 04:21:00 crc kubenswrapper[4770]: I1004 04:21:00.192706 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 04 04:21:00 crc kubenswrapper[4770]: I1004 04:21:00.283616 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.137669 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h59jd"] Oct 04 04:21:04 crc kubenswrapper[4770]: E1004 04:21:04.138467 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="278b89af-8ec2-4258-b7d2-d41e2e58df86" containerName="dnsmasq-dns" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.138488 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="278b89af-8ec2-4258-b7d2-d41e2e58df86" containerName="dnsmasq-dns" Oct 04 04:21:04 crc kubenswrapper[4770]: E1004 04:21:04.138566 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="278b89af-8ec2-4258-b7d2-d41e2e58df86" containerName="init" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.138579 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="278b89af-8ec2-4258-b7d2-d41e2e58df86" containerName="init" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.138846 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="278b89af-8ec2-4258-b7d2-d41e2e58df86" containerName="dnsmasq-dns" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.140614 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.150277 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h59jd"] Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.247190 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/9757975c-8f2e-4e35-853e-b83333f1a8f0-kube-api-access-bnx7x\") pod \"redhat-operators-h59jd\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.247284 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-utilities\") pod \"redhat-operators-h59jd\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.247534 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-catalog-content\") pod \"redhat-operators-h59jd\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.348937 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-catalog-content\") pod \"redhat-operators-h59jd\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.349141 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/9757975c-8f2e-4e35-853e-b83333f1a8f0-kube-api-access-bnx7x\") pod \"redhat-operators-h59jd\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.349184 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-utilities\") pod \"redhat-operators-h59jd\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.349529 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-catalog-content\") pod \"redhat-operators-h59jd\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.349588 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-utilities\") pod \"redhat-operators-h59jd\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.388071 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/9757975c-8f2e-4e35-853e-b83333f1a8f0-kube-api-access-bnx7x\") pod \"redhat-operators-h59jd\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.469317 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.674001 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:21:04 crc kubenswrapper[4770]: E1004 04:21:04.674279 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.799770 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 04 04:21:04 crc kubenswrapper[4770]: I1004 04:21:04.916141 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h59jd"] Oct 04 04:21:04 crc kubenswrapper[4770]: W1004 04:21:04.924222 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9757975c_8f2e_4e35_853e_b83333f1a8f0.slice/crio-481a81cdc484bb571ca1dad3782b01231ca75650796cf4d1286ee3831fa4233e WatchSource:0}: Error finding container 481a81cdc484bb571ca1dad3782b01231ca75650796cf4d1286ee3831fa4233e: Status 404 returned error can't find the container with id 481a81cdc484bb571ca1dad3782b01231ca75650796cf4d1286ee3831fa4233e Oct 04 04:21:05 crc kubenswrapper[4770]: I1004 04:21:05.287360 4770 generic.go:334] "Generic (PLEG): container finished" podID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerID="e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9" exitCode=0 Oct 04 04:21:05 crc kubenswrapper[4770]: I1004 04:21:05.288158 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h59jd" event={"ID":"9757975c-8f2e-4e35-853e-b83333f1a8f0","Type":"ContainerDied","Data":"e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9"} Oct 04 04:21:05 crc kubenswrapper[4770]: I1004 04:21:05.288200 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h59jd" event={"ID":"9757975c-8f2e-4e35-853e-b83333f1a8f0","Type":"ContainerStarted","Data":"481a81cdc484bb571ca1dad3782b01231ca75650796cf4d1286ee3831fa4233e"} Oct 04 04:21:07 crc kubenswrapper[4770]: I1004 04:21:07.307745 4770 generic.go:334] "Generic (PLEG): container finished" podID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerID="5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e" exitCode=0 Oct 04 04:21:07 crc kubenswrapper[4770]: I1004 04:21:07.308294 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h59jd" event={"ID":"9757975c-8f2e-4e35-853e-b83333f1a8f0","Type":"ContainerDied","Data":"5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e"} Oct 04 04:21:08 crc kubenswrapper[4770]: I1004 04:21:08.319269 4770 generic.go:334] "Generic (PLEG): container finished" podID="661cb72b-18eb-47fc-94b3-53fa8c50aec8" containerID="6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c" exitCode=0 Oct 04 04:21:08 crc kubenswrapper[4770]: I1004 04:21:08.319360 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"661cb72b-18eb-47fc-94b3-53fa8c50aec8","Type":"ContainerDied","Data":"6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c"} Oct 04 04:21:08 crc kubenswrapper[4770]: I1004 04:21:08.322196 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h59jd" event={"ID":"9757975c-8f2e-4e35-853e-b83333f1a8f0","Type":"ContainerStarted","Data":"5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f"} Oct 04 04:21:08 crc kubenswrapper[4770]: I1004 04:21:08.379305 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h59jd" podStartSLOduration=1.8359201889999999 podStartE2EDuration="4.379287281s" podCreationTimestamp="2025-10-04 04:21:04 +0000 UTC" firstStartedPulling="2025-10-04 04:21:05.289697883 +0000 UTC m=+4676.581707605" lastFinishedPulling="2025-10-04 04:21:07.833064945 +0000 UTC m=+4679.125074697" observedRunningTime="2025-10-04 04:21:08.377477024 +0000 UTC m=+4679.669486746" watchObservedRunningTime="2025-10-04 04:21:08.379287281 +0000 UTC m=+4679.671296993" Oct 04 04:21:09 crc kubenswrapper[4770]: I1004 04:21:09.332411 4770 generic.go:334] "Generic (PLEG): container finished" podID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" containerID="734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d" exitCode=0 Oct 04 04:21:09 crc kubenswrapper[4770]: I1004 04:21:09.332500 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8f38a480-df1d-4f8c-8009-60f12b7fdb2f","Type":"ContainerDied","Data":"734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d"} Oct 04 04:21:09 crc kubenswrapper[4770]: I1004 04:21:09.335954 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"661cb72b-18eb-47fc-94b3-53fa8c50aec8","Type":"ContainerStarted","Data":"241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3"} Oct 04 04:21:09 crc kubenswrapper[4770]: I1004 04:21:09.336580 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:09 crc kubenswrapper[4770]: I1004 04:21:09.398996 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.398970461 podStartE2EDuration="37.398970461s" podCreationTimestamp="2025-10-04 04:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:21:09.394300729 +0000 UTC m=+4680.686310491" watchObservedRunningTime="2025-10-04 04:21:09.398970461 +0000 UTC m=+4680.690980223" Oct 04 04:21:10 crc kubenswrapper[4770]: I1004 04:21:10.346927 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8f38a480-df1d-4f8c-8009-60f12b7fdb2f","Type":"ContainerStarted","Data":"0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85"} Oct 04 04:21:10 crc kubenswrapper[4770]: I1004 04:21:10.347136 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 04:21:10 crc kubenswrapper[4770]: I1004 04:21:10.385794 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.385770242 podStartE2EDuration="39.385770242s" podCreationTimestamp="2025-10-04 04:20:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:21:10.37574655 +0000 UTC m=+4681.667756252" watchObservedRunningTime="2025-10-04 04:21:10.385770242 +0000 UTC m=+4681.677779954" Oct 04 04:21:14 crc kubenswrapper[4770]: I1004 04:21:14.469645 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:14 crc kubenswrapper[4770]: I1004 04:21:14.470093 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:14 crc kubenswrapper[4770]: I1004 04:21:14.548594 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:15 crc kubenswrapper[4770]: I1004 04:21:15.467553 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:15 crc kubenswrapper[4770]: I1004 04:21:15.540735 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h59jd"] Oct 04 04:21:17 crc kubenswrapper[4770]: I1004 04:21:17.408346 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h59jd" podUID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerName="registry-server" containerID="cri-o://5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f" gracePeriod=2 Oct 04 04:21:17 crc kubenswrapper[4770]: I1004 04:21:17.886122 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:17 crc kubenswrapper[4770]: I1004 04:21:17.961174 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-utilities\") pod \"9757975c-8f2e-4e35-853e-b83333f1a8f0\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " Oct 04 04:21:17 crc kubenswrapper[4770]: I1004 04:21:17.961259 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/9757975c-8f2e-4e35-853e-b83333f1a8f0-kube-api-access-bnx7x\") pod \"9757975c-8f2e-4e35-853e-b83333f1a8f0\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " Oct 04 04:21:17 crc kubenswrapper[4770]: I1004 04:21:17.961474 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-catalog-content\") pod \"9757975c-8f2e-4e35-853e-b83333f1a8f0\" (UID: \"9757975c-8f2e-4e35-853e-b83333f1a8f0\") " Oct 04 04:21:17 crc kubenswrapper[4770]: I1004 04:21:17.962368 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-utilities" (OuterVolumeSpecName: "utilities") pod "9757975c-8f2e-4e35-853e-b83333f1a8f0" (UID: "9757975c-8f2e-4e35-853e-b83333f1a8f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:21:17 crc kubenswrapper[4770]: I1004 04:21:17.973234 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9757975c-8f2e-4e35-853e-b83333f1a8f0-kube-api-access-bnx7x" (OuterVolumeSpecName: "kube-api-access-bnx7x") pod "9757975c-8f2e-4e35-853e-b83333f1a8f0" (UID: "9757975c-8f2e-4e35-853e-b83333f1a8f0"). InnerVolumeSpecName "kube-api-access-bnx7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.063909 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.063956 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnx7x\" (UniqueName: \"kubernetes.io/projected/9757975c-8f2e-4e35-853e-b83333f1a8f0-kube-api-access-bnx7x\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.085110 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9757975c-8f2e-4e35-853e-b83333f1a8f0" (UID: "9757975c-8f2e-4e35-853e-b83333f1a8f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.164979 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9757975c-8f2e-4e35-853e-b83333f1a8f0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.421248 4770 generic.go:334] "Generic (PLEG): container finished" podID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerID="5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f" exitCode=0 Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.421319 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h59jd" event={"ID":"9757975c-8f2e-4e35-853e-b83333f1a8f0","Type":"ContainerDied","Data":"5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f"} Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.421347 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h59jd" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.421366 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h59jd" event={"ID":"9757975c-8f2e-4e35-853e-b83333f1a8f0","Type":"ContainerDied","Data":"481a81cdc484bb571ca1dad3782b01231ca75650796cf4d1286ee3831fa4233e"} Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.421398 4770 scope.go:117] "RemoveContainer" containerID="5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.454414 4770 scope.go:117] "RemoveContainer" containerID="5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.467098 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h59jd"] Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.480626 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h59jd"] Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.486472 4770 scope.go:117] "RemoveContainer" containerID="e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.519817 4770 scope.go:117] "RemoveContainer" containerID="5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f" Oct 04 04:21:18 crc kubenswrapper[4770]: E1004 04:21:18.520340 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f\": container with ID starting with 5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f not found: ID does not exist" containerID="5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.520385 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f"} err="failed to get container status \"5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f\": rpc error: code = NotFound desc = could not find container \"5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f\": container with ID starting with 5413bc02543bdca3c8dabbeb3ac72c968332b4aa96fc0e6d2b05a7c0e570e74f not found: ID does not exist" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.520411 4770 scope.go:117] "RemoveContainer" containerID="5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e" Oct 04 04:21:18 crc kubenswrapper[4770]: E1004 04:21:18.520859 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e\": container with ID starting with 5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e not found: ID does not exist" containerID="5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.520903 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e"} err="failed to get container status \"5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e\": rpc error: code = NotFound desc = could not find container \"5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e\": container with ID starting with 5dee112f1ae96583af220c581e6ace498e938eae357b72f313ff22b86c80673e not found: ID does not exist" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.520931 4770 scope.go:117] "RemoveContainer" containerID="e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9" Oct 04 04:21:18 crc kubenswrapper[4770]: E1004 04:21:18.521290 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9\": container with ID starting with e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9 not found: ID does not exist" containerID="e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.521320 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9"} err="failed to get container status \"e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9\": rpc error: code = NotFound desc = could not find container \"e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9\": container with ID starting with e355fe8678bf0a1d37a7ba24115d976f81f7a33875f121d4fee408d3bcf579b9 not found: ID does not exist" Oct 04 04:21:18 crc kubenswrapper[4770]: I1004 04:21:18.673670 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:21:18 crc kubenswrapper[4770]: E1004 04:21:18.674040 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:21:19 crc kubenswrapper[4770]: I1004 04:21:19.686850 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9757975c-8f2e-4e35-853e-b83333f1a8f0" path="/var/lib/kubelet/pods/9757975c-8f2e-4e35-853e-b83333f1a8f0/volumes" Oct 04 04:21:23 crc kubenswrapper[4770]: I1004 04:21:23.396846 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 04:21:23 crc kubenswrapper[4770]: I1004 04:21:23.501279 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:24 crc kubenswrapper[4770]: I1004 04:21:24.066506 4770 scope.go:117] "RemoveContainer" containerID="3edbbfef4c7a4d5764a7bc1448de8428a060ea348d12236e0a78e4685b3713a9" Oct 04 04:21:24 crc kubenswrapper[4770]: I1004 04:21:24.097390 4770 scope.go:117] "RemoveContainer" containerID="251a11f86c0082eb75c24b6ec96e4d94451362f2e7c734c7673aded7041456cc" Oct 04 04:21:24 crc kubenswrapper[4770]: I1004 04:21:24.123628 4770 scope.go:117] "RemoveContainer" containerID="c0e6e0daed7594b274c32a0e1accc7947b27d7531ccb370a665edbd4c4cfcad9" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.339607 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b6b49df6f-vpzhl"] Oct 04 04:21:30 crc kubenswrapper[4770]: E1004 04:21:30.340567 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerName="registry-server" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.340585 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerName="registry-server" Oct 04 04:21:30 crc kubenswrapper[4770]: E1004 04:21:30.340608 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerName="extract-content" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.340616 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerName="extract-content" Oct 04 04:21:30 crc kubenswrapper[4770]: E1004 04:21:30.340629 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerName="extract-utilities" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.340637 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerName="extract-utilities" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.340830 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9757975c-8f2e-4e35-853e-b83333f1a8f0" containerName="registry-server" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.341659 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.358672 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b6b49df6f-vpzhl"] Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.458184 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-config\") pod \"dnsmasq-dns-7b6b49df6f-vpzhl\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.458467 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-dns-svc\") pod \"dnsmasq-dns-7b6b49df6f-vpzhl\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.459061 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7zkk\" (UniqueName: \"kubernetes.io/projected/89738376-3973-41e0-9ee4-b9157c3c535a-kube-api-access-j7zkk\") pod \"dnsmasq-dns-7b6b49df6f-vpzhl\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.560245 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-config\") pod \"dnsmasq-dns-7b6b49df6f-vpzhl\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.560666 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-dns-svc\") pod \"dnsmasq-dns-7b6b49df6f-vpzhl\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.560778 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7zkk\" (UniqueName: \"kubernetes.io/projected/89738376-3973-41e0-9ee4-b9157c3c535a-kube-api-access-j7zkk\") pod \"dnsmasq-dns-7b6b49df6f-vpzhl\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.561947 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-config\") pod \"dnsmasq-dns-7b6b49df6f-vpzhl\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.562484 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-dns-svc\") pod \"dnsmasq-dns-7b6b49df6f-vpzhl\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.585733 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7zkk\" (UniqueName: \"kubernetes.io/projected/89738376-3973-41e0-9ee4-b9157c3c535a-kube-api-access-j7zkk\") pod \"dnsmasq-dns-7b6b49df6f-vpzhl\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.672990 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.675418 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:21:30 crc kubenswrapper[4770]: E1004 04:21:30.675722 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:21:30 crc kubenswrapper[4770]: I1004 04:21:30.886317 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b6b49df6f-vpzhl"] Oct 04 04:21:31 crc kubenswrapper[4770]: I1004 04:21:31.053638 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:21:31 crc kubenswrapper[4770]: I1004 04:21:31.591647 4770 generic.go:334] "Generic (PLEG): container finished" podID="89738376-3973-41e0-9ee4-b9157c3c535a" containerID="10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a" exitCode=0 Oct 04 04:21:31 crc kubenswrapper[4770]: I1004 04:21:31.591732 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" event={"ID":"89738376-3973-41e0-9ee4-b9157c3c535a","Type":"ContainerDied","Data":"10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a"} Oct 04 04:21:31 crc kubenswrapper[4770]: I1004 04:21:31.591781 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" event={"ID":"89738376-3973-41e0-9ee4-b9157c3c535a","Type":"ContainerStarted","Data":"eddbd90fdad53719497c5f3e710e83fb82a7e8edbb7b1d776050eed6e33d34ff"} Oct 04 04:21:31 crc kubenswrapper[4770]: I1004 04:21:31.810785 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:21:32 crc kubenswrapper[4770]: I1004 04:21:32.600546 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" event={"ID":"89738376-3973-41e0-9ee4-b9157c3c535a","Type":"ContainerStarted","Data":"c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b"} Oct 04 04:21:32 crc kubenswrapper[4770]: I1004 04:21:32.600690 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:32 crc kubenswrapper[4770]: I1004 04:21:32.623318 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" podStartSLOduration=2.623298912 podStartE2EDuration="2.623298912s" podCreationTimestamp="2025-10-04 04:21:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:21:32.619417951 +0000 UTC m=+4703.911427663" watchObservedRunningTime="2025-10-04 04:21:32.623298912 +0000 UTC m=+4703.915308624" Oct 04 04:21:32 crc kubenswrapper[4770]: I1004 04:21:32.978082 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" containerName="rabbitmq" containerID="cri-o://0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85" gracePeriod=604799 Oct 04 04:21:33 crc kubenswrapper[4770]: I1004 04:21:33.394420 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.238:5672: connect: connection refused" Oct 04 04:21:33 crc kubenswrapper[4770]: I1004 04:21:33.603628 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="661cb72b-18eb-47fc-94b3-53fa8c50aec8" containerName="rabbitmq" containerID="cri-o://241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3" gracePeriod=604799 Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.587883 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.660606 4770 generic.go:334] "Generic (PLEG): container finished" podID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" containerID="0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85" exitCode=0 Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.660685 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8f38a480-df1d-4f8c-8009-60f12b7fdb2f","Type":"ContainerDied","Data":"0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85"} Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.660761 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8f38a480-df1d-4f8c-8009-60f12b7fdb2f","Type":"ContainerDied","Data":"af30070c951a239d48949cdfe825688b82ab3001ee4533f340e07b6cca7cd56b"} Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.660815 4770 scope.go:117] "RemoveContainer" containerID="0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.660971 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.694820 4770 scope.go:117] "RemoveContainer" containerID="734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.715386 4770 scope.go:117] "RemoveContainer" containerID="0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85" Oct 04 04:21:39 crc kubenswrapper[4770]: E1004 04:21:39.715806 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85\": container with ID starting with 0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85 not found: ID does not exist" containerID="0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.715857 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85"} err="failed to get container status \"0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85\": rpc error: code = NotFound desc = could not find container \"0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85\": container with ID starting with 0f6a69d627d86c7be59b83ab9b5a086a2631159136f43e14254f069b10c12e85 not found: ID does not exist" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.715903 4770 scope.go:117] "RemoveContainer" containerID="734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d" Oct 04 04:21:39 crc kubenswrapper[4770]: E1004 04:21:39.716790 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d\": container with ID starting with 734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d not found: ID does not exist" containerID="734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.716818 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d"} err="failed to get container status \"734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d\": rpc error: code = NotFound desc = could not find container \"734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d\": container with ID starting with 734d3e7e13cc26dc4b238e38f183211eab8e89ac8e6b74b7580e64a491e22f2d not found: ID does not exist" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.724444 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpgx5\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-kube-api-access-vpgx5\") pod \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.724481 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-plugins\") pod \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.724511 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-server-conf\") pod \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.724610 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") pod \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.724668 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-confd\") pod \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.724705 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-plugins-conf\") pod \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.724731 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-erlang-cookie-secret\") pod \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.724783 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-erlang-cookie\") pod \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.724817 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-pod-info\") pod \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\" (UID: \"8f38a480-df1d-4f8c-8009-60f12b7fdb2f\") " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.727709 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8f38a480-df1d-4f8c-8009-60f12b7fdb2f" (UID: "8f38a480-df1d-4f8c-8009-60f12b7fdb2f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.728355 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8f38a480-df1d-4f8c-8009-60f12b7fdb2f" (UID: "8f38a480-df1d-4f8c-8009-60f12b7fdb2f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.728473 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8f38a480-df1d-4f8c-8009-60f12b7fdb2f" (UID: "8f38a480-df1d-4f8c-8009-60f12b7fdb2f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.731228 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-kube-api-access-vpgx5" (OuterVolumeSpecName: "kube-api-access-vpgx5") pod "8f38a480-df1d-4f8c-8009-60f12b7fdb2f" (UID: "8f38a480-df1d-4f8c-8009-60f12b7fdb2f"). InnerVolumeSpecName "kube-api-access-vpgx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.731274 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8f38a480-df1d-4f8c-8009-60f12b7fdb2f" (UID: "8f38a480-df1d-4f8c-8009-60f12b7fdb2f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.733391 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-pod-info" (OuterVolumeSpecName: "pod-info") pod "8f38a480-df1d-4f8c-8009-60f12b7fdb2f" (UID: "8f38a480-df1d-4f8c-8009-60f12b7fdb2f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.777172 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1" (OuterVolumeSpecName: "persistence") pod "8f38a480-df1d-4f8c-8009-60f12b7fdb2f" (UID: "8f38a480-df1d-4f8c-8009-60f12b7fdb2f"). InnerVolumeSpecName "pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.785480 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-server-conf" (OuterVolumeSpecName: "server-conf") pod "8f38a480-df1d-4f8c-8009-60f12b7fdb2f" (UID: "8f38a480-df1d-4f8c-8009-60f12b7fdb2f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.826318 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.826668 4770 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.826774 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpgx5\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-kube-api-access-vpgx5\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.826842 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.826926 4770 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.827110 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") on node \"crc\" " Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.827221 4770 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.827347 4770 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.826975 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8f38a480-df1d-4f8c-8009-60f12b7fdb2f" (UID: "8f38a480-df1d-4f8c-8009-60f12b7fdb2f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.841817 4770 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.841947 4770 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1") on node "crc" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.928271 4770 reconciler_common.go:293] "Volume detached for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:39 crc kubenswrapper[4770]: I1004 04:21:39.928310 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f38a480-df1d-4f8c-8009-60f12b7fdb2f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.008167 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.025135 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.038992 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:21:40 crc kubenswrapper[4770]: E1004 04:21:40.039313 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" containerName="rabbitmq" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.039331 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" containerName="rabbitmq" Oct 04 04:21:40 crc kubenswrapper[4770]: E1004 04:21:40.039354 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" containerName="setup-container" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.039361 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" containerName="setup-container" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.039511 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" containerName="rabbitmq" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.040265 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.042329 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.042352 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.042541 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.042598 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6z9fp" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.043097 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.049943 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.134484 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.134530 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.134555 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.134585 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.134601 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.134654 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.134669 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.134692 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.134710 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwhs5\" (UniqueName: \"kubernetes.io/projected/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-kube-api-access-dwhs5\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.151666 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236029 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") pod \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236078 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-plugins-conf\") pod \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236145 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-server-conf\") pod \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236186 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-plugins\") pod \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236232 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-confd\") pod \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236275 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/661cb72b-18eb-47fc-94b3-53fa8c50aec8-pod-info\") pod \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236298 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvgbn\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-kube-api-access-hvgbn\") pod \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236322 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/661cb72b-18eb-47fc-94b3-53fa8c50aec8-erlang-cookie-secret\") pod \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236363 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-erlang-cookie\") pod \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\" (UID: \"661cb72b-18eb-47fc-94b3-53fa8c50aec8\") " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236570 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236602 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236626 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236653 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236670 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236730 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236750 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236772 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.236794 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwhs5\" (UniqueName: \"kubernetes.io/projected/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-kube-api-access-dwhs5\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.237706 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "661cb72b-18eb-47fc-94b3-53fa8c50aec8" (UID: "661cb72b-18eb-47fc-94b3-53fa8c50aec8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.238349 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.238725 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.239010 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.240040 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.242203 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.238733 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "661cb72b-18eb-47fc-94b3-53fa8c50aec8" (UID: "661cb72b-18eb-47fc-94b3-53fa8c50aec8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.243281 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "661cb72b-18eb-47fc-94b3-53fa8c50aec8" (UID: "661cb72b-18eb-47fc-94b3-53fa8c50aec8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.245167 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/661cb72b-18eb-47fc-94b3-53fa8c50aec8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "661cb72b-18eb-47fc-94b3-53fa8c50aec8" (UID: "661cb72b-18eb-47fc-94b3-53fa8c50aec8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.246143 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.246529 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-kube-api-access-hvgbn" (OuterVolumeSpecName: "kube-api-access-hvgbn") pod "661cb72b-18eb-47fc-94b3-53fa8c50aec8" (UID: "661cb72b-18eb-47fc-94b3-53fa8c50aec8"). InnerVolumeSpecName "kube-api-access-hvgbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.248347 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.248382 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/906571f9b5ef483062810d0a89b78f264350f028b315b995e454b2cd8a5402c8/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.248515 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.254761 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/661cb72b-18eb-47fc-94b3-53fa8c50aec8-pod-info" (OuterVolumeSpecName: "pod-info") pod "661cb72b-18eb-47fc-94b3-53fa8c50aec8" (UID: "661cb72b-18eb-47fc-94b3-53fa8c50aec8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.258864 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2" (OuterVolumeSpecName: "persistence") pod "661cb72b-18eb-47fc-94b3-53fa8c50aec8" (UID: "661cb72b-18eb-47fc-94b3-53fa8c50aec8"). InnerVolumeSpecName "pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.259379 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwhs5\" (UniqueName: \"kubernetes.io/projected/b8671f84-8a47-4d20-8272-90b7fd4d6b3d-kube-api-access-dwhs5\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.265137 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-server-conf" (OuterVolumeSpecName: "server-conf") pod "661cb72b-18eb-47fc-94b3-53fa8c50aec8" (UID: "661cb72b-18eb-47fc-94b3-53fa8c50aec8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.290575 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5b1d489-1882-4b30-bc6f-856ca83feed1\") pod \"rabbitmq-server-0\" (UID: \"b8671f84-8a47-4d20-8272-90b7fd4d6b3d\") " pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.311078 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "661cb72b-18eb-47fc-94b3-53fa8c50aec8" (UID: "661cb72b-18eb-47fc-94b3-53fa8c50aec8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.338200 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.338229 4770 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/661cb72b-18eb-47fc-94b3-53fa8c50aec8-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.338239 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvgbn\" (UniqueName: \"kubernetes.io/projected/661cb72b-18eb-47fc-94b3-53fa8c50aec8-kube-api-access-hvgbn\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.338252 4770 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/661cb72b-18eb-47fc-94b3-53fa8c50aec8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.338260 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.338294 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") on node \"crc\" " Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.338303 4770 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.338313 4770 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/661cb72b-18eb-47fc-94b3-53fa8c50aec8-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.338321 4770 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/661cb72b-18eb-47fc-94b3-53fa8c50aec8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.362669 4770 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.362809 4770 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2") on node "crc" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.365141 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.439217 4770 reconciler_common.go:293] "Volume detached for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.670850 4770 generic.go:334] "Generic (PLEG): container finished" podID="661cb72b-18eb-47fc-94b3-53fa8c50aec8" containerID="241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3" exitCode=0 Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.670922 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"661cb72b-18eb-47fc-94b3-53fa8c50aec8","Type":"ContainerDied","Data":"241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3"} Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.670968 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"661cb72b-18eb-47fc-94b3-53fa8c50aec8","Type":"ContainerDied","Data":"e9874c42f82487ac92586e4d2dd486408067a0d7d1da1bdc4e66dd7fdfb17d39"} Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.670986 4770 scope.go:117] "RemoveContainer" containerID="241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.671225 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.675737 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.706700 4770 scope.go:117] "RemoveContainer" containerID="6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.735030 4770 scope.go:117] "RemoveContainer" containerID="241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3" Oct 04 04:21:40 crc kubenswrapper[4770]: E1004 04:21:40.744712 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3\": container with ID starting with 241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3 not found: ID does not exist" containerID="241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.744762 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3"} err="failed to get container status \"241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3\": rpc error: code = NotFound desc = could not find container \"241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3\": container with ID starting with 241402df4e0cfe6990828b52005a4e231045366ee337da8cc43eca378692d1a3 not found: ID does not exist" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.744795 4770 scope.go:117] "RemoveContainer" containerID="6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c" Oct 04 04:21:40 crc kubenswrapper[4770]: E1004 04:21:40.750040 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c\": container with ID starting with 6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c not found: ID does not exist" containerID="6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.757213 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c"} err="failed to get container status \"6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c\": rpc error: code = NotFound desc = could not find container \"6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c\": container with ID starting with 6737ee47f3ebf4ad016156951d592cef67bbd3e8e8d986a0b2dbda5ac9dc728c not found: ID does not exist" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.757284 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.766399 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.780233 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64547bf95c-lm8q9"] Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.780452 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" podUID="5c978011-2d08-4128-a442-4f4e59ff7587" containerName="dnsmasq-dns" containerID="cri-o://64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896" gracePeriod=10 Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.791524 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:21:40 crc kubenswrapper[4770]: E1004 04:21:40.791902 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661cb72b-18eb-47fc-94b3-53fa8c50aec8" containerName="rabbitmq" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.791919 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="661cb72b-18eb-47fc-94b3-53fa8c50aec8" containerName="rabbitmq" Oct 04 04:21:40 crc kubenswrapper[4770]: E1004 04:21:40.791935 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661cb72b-18eb-47fc-94b3-53fa8c50aec8" containerName="setup-container" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.791943 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="661cb72b-18eb-47fc-94b3-53fa8c50aec8" containerName="setup-container" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.792233 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="661cb72b-18eb-47fc-94b3-53fa8c50aec8" containerName="rabbitmq" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.793253 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.796382 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.796562 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.796749 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.797035 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qlcz4" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.797153 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.801230 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.805562 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.849105 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/60cb9246-1987-4782-97e3-84935639077f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.849198 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/60cb9246-1987-4782-97e3-84935639077f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.849235 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/60cb9246-1987-4782-97e3-84935639077f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.849293 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7csf\" (UniqueName: \"kubernetes.io/projected/60cb9246-1987-4782-97e3-84935639077f-kube-api-access-d7csf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.849316 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/60cb9246-1987-4782-97e3-84935639077f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.849375 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/60cb9246-1987-4782-97e3-84935639077f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.849582 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/60cb9246-1987-4782-97e3-84935639077f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.849873 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/60cb9246-1987-4782-97e3-84935639077f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.849909 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.951634 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/60cb9246-1987-4782-97e3-84935639077f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.951675 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.951721 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/60cb9246-1987-4782-97e3-84935639077f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.951751 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/60cb9246-1987-4782-97e3-84935639077f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.951779 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/60cb9246-1987-4782-97e3-84935639077f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.951804 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7csf\" (UniqueName: \"kubernetes.io/projected/60cb9246-1987-4782-97e3-84935639077f-kube-api-access-d7csf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.951824 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/60cb9246-1987-4782-97e3-84935639077f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.951853 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/60cb9246-1987-4782-97e3-84935639077f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.951868 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/60cb9246-1987-4782-97e3-84935639077f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.952340 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/60cb9246-1987-4782-97e3-84935639077f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.952349 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/60cb9246-1987-4782-97e3-84935639077f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.953716 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/60cb9246-1987-4782-97e3-84935639077f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.955912 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/60cb9246-1987-4782-97e3-84935639077f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.956621 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/60cb9246-1987-4782-97e3-84935639077f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.956674 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/60cb9246-1987-4782-97e3-84935639077f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.958202 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/60cb9246-1987-4782-97e3-84935639077f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.958310 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.958335 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/be312a493dc81d83360396620ad55cfc993563ccce411951daf7b4e05473c6be/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.970848 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7csf\" (UniqueName: \"kubernetes.io/projected/60cb9246-1987-4782-97e3-84935639077f-kube-api-access-d7csf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:40 crc kubenswrapper[4770]: I1004 04:21:40.995529 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25a07b22-5115-47f4-b0e5-6c8bae6478c2\") pod \"rabbitmq-cell1-server-0\" (UID: \"60cb9246-1987-4782-97e3-84935639077f\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.101711 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.122312 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.154860 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-config\") pod \"5c978011-2d08-4128-a442-4f4e59ff7587\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.154940 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg4m8\" (UniqueName: \"kubernetes.io/projected/5c978011-2d08-4128-a442-4f4e59ff7587-kube-api-access-dg4m8\") pod \"5c978011-2d08-4128-a442-4f4e59ff7587\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.154969 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-dns-svc\") pod \"5c978011-2d08-4128-a442-4f4e59ff7587\" (UID: \"5c978011-2d08-4128-a442-4f4e59ff7587\") " Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.158773 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c978011-2d08-4128-a442-4f4e59ff7587-kube-api-access-dg4m8" (OuterVolumeSpecName: "kube-api-access-dg4m8") pod "5c978011-2d08-4128-a442-4f4e59ff7587" (UID: "5c978011-2d08-4128-a442-4f4e59ff7587"). InnerVolumeSpecName "kube-api-access-dg4m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.187074 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-config" (OuterVolumeSpecName: "config") pod "5c978011-2d08-4128-a442-4f4e59ff7587" (UID: "5c978011-2d08-4128-a442-4f4e59ff7587"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.187198 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5c978011-2d08-4128-a442-4f4e59ff7587" (UID: "5c978011-2d08-4128-a442-4f4e59ff7587"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.256511 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.256546 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg4m8\" (UniqueName: \"kubernetes.io/projected/5c978011-2d08-4128-a442-4f4e59ff7587-kube-api-access-dg4m8\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.256557 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c978011-2d08-4128-a442-4f4e59ff7587-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.574505 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:21:41 crc kubenswrapper[4770]: W1004 04:21:41.606448 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60cb9246_1987_4782_97e3_84935639077f.slice/crio-5f08a6ca81b5454c766581b8a33f41f2ecfe65465e3a65c89d7fad3fcda49a53 WatchSource:0}: Error finding container 5f08a6ca81b5454c766581b8a33f41f2ecfe65465e3a65c89d7fad3fcda49a53: Status 404 returned error can't find the container with id 5f08a6ca81b5454c766581b8a33f41f2ecfe65465e3a65c89d7fad3fcda49a53 Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.693899 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="661cb72b-18eb-47fc-94b3-53fa8c50aec8" path="/var/lib/kubelet/pods/661cb72b-18eb-47fc-94b3-53fa8c50aec8/volumes" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.695217 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f38a480-df1d-4f8c-8009-60f12b7fdb2f" path="/var/lib/kubelet/pods/8f38a480-df1d-4f8c-8009-60f12b7fdb2f/volumes" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.698553 4770 generic.go:334] "Generic (PLEG): container finished" podID="5c978011-2d08-4128-a442-4f4e59ff7587" containerID="64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896" exitCode=0 Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.698624 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" event={"ID":"5c978011-2d08-4128-a442-4f4e59ff7587","Type":"ContainerDied","Data":"64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896"} Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.698669 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.698780 4770 scope.go:117] "RemoveContainer" containerID="64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.698884 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64547bf95c-lm8q9" event={"ID":"5c978011-2d08-4128-a442-4f4e59ff7587","Type":"ContainerDied","Data":"755178700ca8eb8ec48c5d580d7149abb2e97d2639d6f88832fb902f743bb035"} Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.700038 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"60cb9246-1987-4782-97e3-84935639077f","Type":"ContainerStarted","Data":"5f08a6ca81b5454c766581b8a33f41f2ecfe65465e3a65c89d7fad3fcda49a53"} Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.703792 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8671f84-8a47-4d20-8272-90b7fd4d6b3d","Type":"ContainerStarted","Data":"e5924dfab7aaf2bfd8a9e6ade23ac019e5023b33bb6403eb2ed7d8ce92c14246"} Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.737491 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64547bf95c-lm8q9"] Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.742165 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64547bf95c-lm8q9"] Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.757417 4770 scope.go:117] "RemoveContainer" containerID="8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.780421 4770 scope.go:117] "RemoveContainer" containerID="64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896" Oct 04 04:21:41 crc kubenswrapper[4770]: E1004 04:21:41.781061 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896\": container with ID starting with 64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896 not found: ID does not exist" containerID="64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.781106 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896"} err="failed to get container status \"64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896\": rpc error: code = NotFound desc = could not find container \"64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896\": container with ID starting with 64c1ca8f0dbf08468ac7360f3327e64f6aa4d6e5461d0599f54f549719238896 not found: ID does not exist" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.781136 4770 scope.go:117] "RemoveContainer" containerID="8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac" Oct 04 04:21:41 crc kubenswrapper[4770]: E1004 04:21:41.781662 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac\": container with ID starting with 8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac not found: ID does not exist" containerID="8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac" Oct 04 04:21:41 crc kubenswrapper[4770]: I1004 04:21:41.781696 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac"} err="failed to get container status \"8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac\": rpc error: code = NotFound desc = could not find container \"8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac\": container with ID starting with 8fd2c5c2ff19f74ae7d456d4fcb83d59091074de0500b666b6c75daa1417d6ac not found: ID does not exist" Oct 04 04:21:42 crc kubenswrapper[4770]: I1004 04:21:42.715477 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8671f84-8a47-4d20-8272-90b7fd4d6b3d","Type":"ContainerStarted","Data":"7fb8bfc302c1efd6fd2bc093cc8738f0a670385866aaa308208b43d5f1d41dde"} Oct 04 04:21:43 crc kubenswrapper[4770]: I1004 04:21:43.674637 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:21:43 crc kubenswrapper[4770]: E1004 04:21:43.675276 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:21:43 crc kubenswrapper[4770]: I1004 04:21:43.690779 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c978011-2d08-4128-a442-4f4e59ff7587" path="/var/lib/kubelet/pods/5c978011-2d08-4128-a442-4f4e59ff7587/volumes" Oct 04 04:21:43 crc kubenswrapper[4770]: I1004 04:21:43.729309 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"60cb9246-1987-4782-97e3-84935639077f","Type":"ContainerStarted","Data":"389c438790c21d8bb90c83f5bac915445d8ce67b7e593bee55a1723e53716e69"} Oct 04 04:21:58 crc kubenswrapper[4770]: I1004 04:21:58.673751 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:21:58 crc kubenswrapper[4770]: E1004 04:21:58.674569 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:22:12 crc kubenswrapper[4770]: I1004 04:22:12.673804 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:22:12 crc kubenswrapper[4770]: E1004 04:22:12.674601 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:22:16 crc kubenswrapper[4770]: I1004 04:22:16.020169 4770 generic.go:334] "Generic (PLEG): container finished" podID="b8671f84-8a47-4d20-8272-90b7fd4d6b3d" containerID="7fb8bfc302c1efd6fd2bc093cc8738f0a670385866aaa308208b43d5f1d41dde" exitCode=0 Oct 04 04:22:16 crc kubenswrapper[4770]: I1004 04:22:16.020276 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8671f84-8a47-4d20-8272-90b7fd4d6b3d","Type":"ContainerDied","Data":"7fb8bfc302c1efd6fd2bc093cc8738f0a670385866aaa308208b43d5f1d41dde"} Oct 04 04:22:17 crc kubenswrapper[4770]: I1004 04:22:17.029076 4770 generic.go:334] "Generic (PLEG): container finished" podID="60cb9246-1987-4782-97e3-84935639077f" containerID="389c438790c21d8bb90c83f5bac915445d8ce67b7e593bee55a1723e53716e69" exitCode=0 Oct 04 04:22:17 crc kubenswrapper[4770]: I1004 04:22:17.029182 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"60cb9246-1987-4782-97e3-84935639077f","Type":"ContainerDied","Data":"389c438790c21d8bb90c83f5bac915445d8ce67b7e593bee55a1723e53716e69"} Oct 04 04:22:17 crc kubenswrapper[4770]: I1004 04:22:17.031397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b8671f84-8a47-4d20-8272-90b7fd4d6b3d","Type":"ContainerStarted","Data":"95a74f1251bc62ed29580862c6fb99c1dcaa70c3d5843c02f338522e59582611"} Oct 04 04:22:17 crc kubenswrapper[4770]: I1004 04:22:17.031644 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 04:22:17 crc kubenswrapper[4770]: I1004 04:22:17.129591 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.129574235 podStartE2EDuration="38.129574235s" podCreationTimestamp="2025-10-04 04:21:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:22:17.121531365 +0000 UTC m=+4748.413541087" watchObservedRunningTime="2025-10-04 04:22:17.129574235 +0000 UTC m=+4748.421583947" Oct 04 04:22:18 crc kubenswrapper[4770]: I1004 04:22:18.042228 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"60cb9246-1987-4782-97e3-84935639077f","Type":"ContainerStarted","Data":"c6a974fea812abc6a8004581d0d919091662a77807fbaab57e1c881e37b5728b"} Oct 04 04:22:18 crc kubenswrapper[4770]: I1004 04:22:18.042786 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:22:23 crc kubenswrapper[4770]: I1004 04:22:23.674173 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:22:23 crc kubenswrapper[4770]: E1004 04:22:23.674970 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:22:30 crc kubenswrapper[4770]: I1004 04:22:30.367684 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b8671f84-8a47-4d20-8272-90b7fd4d6b3d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.245:5672: connect: connection refused" Oct 04 04:22:31 crc kubenswrapper[4770]: I1004 04:22:31.124257 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="60cb9246-1987-4782-97e3-84935639077f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.246:5672: connect: connection refused" Oct 04 04:22:38 crc kubenswrapper[4770]: I1004 04:22:38.673863 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:22:38 crc kubenswrapper[4770]: E1004 04:22:38.674665 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:22:40 crc kubenswrapper[4770]: I1004 04:22:40.367077 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b8671f84-8a47-4d20-8272-90b7fd4d6b3d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.245:5672: connect: connection refused" Oct 04 04:22:41 crc kubenswrapper[4770]: I1004 04:22:41.123500 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="60cb9246-1987-4782-97e3-84935639077f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.246:5672: connect: connection refused" Oct 04 04:22:50 crc kubenswrapper[4770]: I1004 04:22:50.368080 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 04:22:50 crc kubenswrapper[4770]: I1004 04:22:50.408053 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=70.407997759 podStartE2EDuration="1m10.407997759s" podCreationTimestamp="2025-10-04 04:21:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:22:18.082600894 +0000 UTC m=+4749.374610666" watchObservedRunningTime="2025-10-04 04:22:50.407997759 +0000 UTC m=+4781.700007491" Oct 04 04:22:51 crc kubenswrapper[4770]: I1004 04:22:51.124397 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:22:53 crc kubenswrapper[4770]: I1004 04:22:53.673766 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:22:53 crc kubenswrapper[4770]: E1004 04:22:53.674679 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.094958 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 04 04:22:59 crc kubenswrapper[4770]: E1004 04:22:59.095780 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c978011-2d08-4128-a442-4f4e59ff7587" containerName="init" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.095803 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c978011-2d08-4128-a442-4f4e59ff7587" containerName="init" Oct 04 04:22:59 crc kubenswrapper[4770]: E1004 04:22:59.095823 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c978011-2d08-4128-a442-4f4e59ff7587" containerName="dnsmasq-dns" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.095835 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c978011-2d08-4128-a442-4f4e59ff7587" containerName="dnsmasq-dns" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.096183 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c978011-2d08-4128-a442-4f4e59ff7587" containerName="dnsmasq-dns" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.097463 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.113682 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t7x2l" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.141855 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.199076 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq2dq\" (UniqueName: \"kubernetes.io/projected/c624ee24-c71e-46d3-9510-0dfc458eb46b-kube-api-access-bq2dq\") pod \"mariadb-client-1-default\" (UID: \"c624ee24-c71e-46d3-9510-0dfc458eb46b\") " pod="openstack/mariadb-client-1-default" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.300904 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq2dq\" (UniqueName: \"kubernetes.io/projected/c624ee24-c71e-46d3-9510-0dfc458eb46b-kube-api-access-bq2dq\") pod \"mariadb-client-1-default\" (UID: \"c624ee24-c71e-46d3-9510-0dfc458eb46b\") " pod="openstack/mariadb-client-1-default" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.333684 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq2dq\" (UniqueName: \"kubernetes.io/projected/c624ee24-c71e-46d3-9510-0dfc458eb46b-kube-api-access-bq2dq\") pod \"mariadb-client-1-default\" (UID: \"c624ee24-c71e-46d3-9510-0dfc458eb46b\") " pod="openstack/mariadb-client-1-default" Oct 04 04:22:59 crc kubenswrapper[4770]: I1004 04:22:59.451549 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 04 04:23:00 crc kubenswrapper[4770]: I1004 04:23:00.033994 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 04 04:23:00 crc kubenswrapper[4770]: W1004 04:23:00.040477 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc624ee24_c71e_46d3_9510_0dfc458eb46b.slice/crio-89d544dda639f244940728a5c25bc0698d738c4870190e9b36fcd0e4c049732e WatchSource:0}: Error finding container 89d544dda639f244940728a5c25bc0698d738c4870190e9b36fcd0e4c049732e: Status 404 returned error can't find the container with id 89d544dda639f244940728a5c25bc0698d738c4870190e9b36fcd0e4c049732e Oct 04 04:23:00 crc kubenswrapper[4770]: I1004 04:23:00.043165 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:23:00 crc kubenswrapper[4770]: I1004 04:23:00.440896 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"c624ee24-c71e-46d3-9510-0dfc458eb46b","Type":"ContainerStarted","Data":"89d544dda639f244940728a5c25bc0698d738c4870190e9b36fcd0e4c049732e"} Oct 04 04:23:02 crc kubenswrapper[4770]: I1004 04:23:02.460198 4770 generic.go:334] "Generic (PLEG): container finished" podID="c624ee24-c71e-46d3-9510-0dfc458eb46b" containerID="0380ddfb1198e43d4716d1111bd20230db09c0b5eded5ec6826d0222a67ddd86" exitCode=0 Oct 04 04:23:02 crc kubenswrapper[4770]: I1004 04:23:02.460251 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"c624ee24-c71e-46d3-9510-0dfc458eb46b","Type":"ContainerDied","Data":"0380ddfb1198e43d4716d1111bd20230db09c0b5eded5ec6826d0222a67ddd86"} Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.020502 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.051672 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_c624ee24-c71e-46d3-9510-0dfc458eb46b/mariadb-client-1-default/0.log" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.074123 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq2dq\" (UniqueName: \"kubernetes.io/projected/c624ee24-c71e-46d3-9510-0dfc458eb46b-kube-api-access-bq2dq\") pod \"c624ee24-c71e-46d3-9510-0dfc458eb46b\" (UID: \"c624ee24-c71e-46d3-9510-0dfc458eb46b\") " Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.086243 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c624ee24-c71e-46d3-9510-0dfc458eb46b-kube-api-access-bq2dq" (OuterVolumeSpecName: "kube-api-access-bq2dq") pod "c624ee24-c71e-46d3-9510-0dfc458eb46b" (UID: "c624ee24-c71e-46d3-9510-0dfc458eb46b"). InnerVolumeSpecName "kube-api-access-bq2dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.089079 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.095793 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.176309 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq2dq\" (UniqueName: \"kubernetes.io/projected/c624ee24-c71e-46d3-9510-0dfc458eb46b-kube-api-access-bq2dq\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.490066 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89d544dda639f244940728a5c25bc0698d738c4870190e9b36fcd0e4c049732e" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.490131 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.667116 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 04 04:23:04 crc kubenswrapper[4770]: E1004 04:23:04.667593 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c624ee24-c71e-46d3-9510-0dfc458eb46b" containerName="mariadb-client-1-default" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.667627 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c624ee24-c71e-46d3-9510-0dfc458eb46b" containerName="mariadb-client-1-default" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.667889 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c624ee24-c71e-46d3-9510-0dfc458eb46b" containerName="mariadb-client-1-default" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.668664 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.672865 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t7x2l" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.674851 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.786086 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z56kv\" (UniqueName: \"kubernetes.io/projected/56e6816e-bb69-4485-b532-b2624f39400a-kube-api-access-z56kv\") pod \"mariadb-client-2-default\" (UID: \"56e6816e-bb69-4485-b532-b2624f39400a\") " pod="openstack/mariadb-client-2-default" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.887858 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z56kv\" (UniqueName: \"kubernetes.io/projected/56e6816e-bb69-4485-b532-b2624f39400a-kube-api-access-z56kv\") pod \"mariadb-client-2-default\" (UID: \"56e6816e-bb69-4485-b532-b2624f39400a\") " pod="openstack/mariadb-client-2-default" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.910398 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z56kv\" (UniqueName: \"kubernetes.io/projected/56e6816e-bb69-4485-b532-b2624f39400a-kube-api-access-z56kv\") pod \"mariadb-client-2-default\" (UID: \"56e6816e-bb69-4485-b532-b2624f39400a\") " pod="openstack/mariadb-client-2-default" Oct 04 04:23:04 crc kubenswrapper[4770]: I1004 04:23:04.995197 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 04 04:23:05 crc kubenswrapper[4770]: I1004 04:23:05.541854 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 04 04:23:05 crc kubenswrapper[4770]: W1004 04:23:05.548674 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56e6816e_bb69_4485_b532_b2624f39400a.slice/crio-fc7d01f6ee066d099204372622131d56176a26e02412aaf52c79fb5ee75e44ec WatchSource:0}: Error finding container fc7d01f6ee066d099204372622131d56176a26e02412aaf52c79fb5ee75e44ec: Status 404 returned error can't find the container with id fc7d01f6ee066d099204372622131d56176a26e02412aaf52c79fb5ee75e44ec Oct 04 04:23:05 crc kubenswrapper[4770]: I1004 04:23:05.685956 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c624ee24-c71e-46d3-9510-0dfc458eb46b" path="/var/lib/kubelet/pods/c624ee24-c71e-46d3-9510-0dfc458eb46b/volumes" Oct 04 04:23:06 crc kubenswrapper[4770]: I1004 04:23:06.510106 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"56e6816e-bb69-4485-b532-b2624f39400a","Type":"ContainerStarted","Data":"8191e38f4305ebb77c06af49baee1b27071748575b5c7db6966479d90b8e534e"} Oct 04 04:23:06 crc kubenswrapper[4770]: I1004 04:23:06.510689 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"56e6816e-bb69-4485-b532-b2624f39400a","Type":"ContainerStarted","Data":"fc7d01f6ee066d099204372622131d56176a26e02412aaf52c79fb5ee75e44ec"} Oct 04 04:23:06 crc kubenswrapper[4770]: I1004 04:23:06.539354 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=2.539335369 podStartE2EDuration="2.539335369s" podCreationTimestamp="2025-10-04 04:23:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:23:06.527241293 +0000 UTC m=+4797.819251025" watchObservedRunningTime="2025-10-04 04:23:06.539335369 +0000 UTC m=+4797.831345081" Oct 04 04:23:06 crc kubenswrapper[4770]: I1004 04:23:06.673420 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:23:06 crc kubenswrapper[4770]: E1004 04:23:06.673639 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:23:07 crc kubenswrapper[4770]: I1004 04:23:07.520349 4770 generic.go:334] "Generic (PLEG): container finished" podID="56e6816e-bb69-4485-b532-b2624f39400a" containerID="8191e38f4305ebb77c06af49baee1b27071748575b5c7db6966479d90b8e534e" exitCode=0 Oct 04 04:23:07 crc kubenswrapper[4770]: I1004 04:23:07.520430 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"56e6816e-bb69-4485-b532-b2624f39400a","Type":"ContainerDied","Data":"8191e38f4305ebb77c06af49baee1b27071748575b5c7db6966479d90b8e534e"} Oct 04 04:23:08 crc kubenswrapper[4770]: I1004 04:23:08.861474 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 04 04:23:08 crc kubenswrapper[4770]: I1004 04:23:08.899255 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 04 04:23:08 crc kubenswrapper[4770]: I1004 04:23:08.905318 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 04 04:23:08 crc kubenswrapper[4770]: I1004 04:23:08.953828 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z56kv\" (UniqueName: \"kubernetes.io/projected/56e6816e-bb69-4485-b532-b2624f39400a-kube-api-access-z56kv\") pod \"56e6816e-bb69-4485-b532-b2624f39400a\" (UID: \"56e6816e-bb69-4485-b532-b2624f39400a\") " Oct 04 04:23:08 crc kubenswrapper[4770]: I1004 04:23:08.959122 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56e6816e-bb69-4485-b532-b2624f39400a-kube-api-access-z56kv" (OuterVolumeSpecName: "kube-api-access-z56kv") pod "56e6816e-bb69-4485-b532-b2624f39400a" (UID: "56e6816e-bb69-4485-b532-b2624f39400a"). InnerVolumeSpecName "kube-api-access-z56kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.055951 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z56kv\" (UniqueName: \"kubernetes.io/projected/56e6816e-bb69-4485-b532-b2624f39400a-kube-api-access-z56kv\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.429280 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 04 04:23:09 crc kubenswrapper[4770]: E1004 04:23:09.430673 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e6816e-bb69-4485-b532-b2624f39400a" containerName="mariadb-client-2-default" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.430722 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e6816e-bb69-4485-b532-b2624f39400a" containerName="mariadb-client-2-default" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.431143 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e6816e-bb69-4485-b532-b2624f39400a" containerName="mariadb-client-2-default" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.432314 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.449810 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.460175 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwtdc\" (UniqueName: \"kubernetes.io/projected/957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5-kube-api-access-xwtdc\") pod \"mariadb-client-1\" (UID: \"957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5\") " pod="openstack/mariadb-client-1" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.536267 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc7d01f6ee066d099204372622131d56176a26e02412aaf52c79fb5ee75e44ec" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.536332 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.561410 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwtdc\" (UniqueName: \"kubernetes.io/projected/957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5-kube-api-access-xwtdc\") pod \"mariadb-client-1\" (UID: \"957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5\") " pod="openstack/mariadb-client-1" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.581046 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwtdc\" (UniqueName: \"kubernetes.io/projected/957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5-kube-api-access-xwtdc\") pod \"mariadb-client-1\" (UID: \"957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5\") " pod="openstack/mariadb-client-1" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.687239 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56e6816e-bb69-4485-b532-b2624f39400a" path="/var/lib/kubelet/pods/56e6816e-bb69-4485-b532-b2624f39400a/volumes" Oct 04 04:23:09 crc kubenswrapper[4770]: I1004 04:23:09.772746 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 04 04:23:11 crc kubenswrapper[4770]: I1004 04:23:11.535490 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 04 04:23:11 crc kubenswrapper[4770]: I1004 04:23:11.555735 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5","Type":"ContainerStarted","Data":"61d32c9bcae30fc6c08878b884552a7fc7b4a2cb05d5ee51a8e0fcd8fc0d99d4"} Oct 04 04:23:12 crc kubenswrapper[4770]: I1004 04:23:12.574394 4770 generic.go:334] "Generic (PLEG): container finished" podID="957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5" containerID="99d9238982da4cf11dc900c20fff1296210edfd18f43b37e905e5e5d354f4ebd" exitCode=0 Oct 04 04:23:12 crc kubenswrapper[4770]: I1004 04:23:12.574531 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5","Type":"ContainerDied","Data":"99d9238982da4cf11dc900c20fff1296210edfd18f43b37e905e5e5d354f4ebd"} Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.057232 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.078334 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5/mariadb-client-1/0.log" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.109624 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.115266 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.139515 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwtdc\" (UniqueName: \"kubernetes.io/projected/957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5-kube-api-access-xwtdc\") pod \"957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5\" (UID: \"957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5\") " Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.146105 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5-kube-api-access-xwtdc" (OuterVolumeSpecName: "kube-api-access-xwtdc") pod "957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5" (UID: "957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5"). InnerVolumeSpecName "kube-api-access-xwtdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.241441 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwtdc\" (UniqueName: \"kubernetes.io/projected/957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5-kube-api-access-xwtdc\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.594968 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61d32c9bcae30fc6c08878b884552a7fc7b4a2cb05d5ee51a8e0fcd8fc0d99d4" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.595031 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.653114 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 04 04:23:14 crc kubenswrapper[4770]: E1004 04:23:14.653582 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5" containerName="mariadb-client-1" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.653605 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5" containerName="mariadb-client-1" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.653992 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5" containerName="mariadb-client-1" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.654724 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.658293 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t7x2l" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.664954 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.749278 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mz5g\" (UniqueName: \"kubernetes.io/projected/5a44192f-c876-4ef8-835a-ee99845c613e-kube-api-access-2mz5g\") pod \"mariadb-client-4-default\" (UID: \"5a44192f-c876-4ef8-835a-ee99845c613e\") " pod="openstack/mariadb-client-4-default" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.850913 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mz5g\" (UniqueName: \"kubernetes.io/projected/5a44192f-c876-4ef8-835a-ee99845c613e-kube-api-access-2mz5g\") pod \"mariadb-client-4-default\" (UID: \"5a44192f-c876-4ef8-835a-ee99845c613e\") " pod="openstack/mariadb-client-4-default" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.869844 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mz5g\" (UniqueName: \"kubernetes.io/projected/5a44192f-c876-4ef8-835a-ee99845c613e-kube-api-access-2mz5g\") pod \"mariadb-client-4-default\" (UID: \"5a44192f-c876-4ef8-835a-ee99845c613e\") " pod="openstack/mariadb-client-4-default" Oct 04 04:23:14 crc kubenswrapper[4770]: I1004 04:23:14.989340 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 04 04:23:15 crc kubenswrapper[4770]: I1004 04:23:15.558553 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 04 04:23:15 crc kubenswrapper[4770]: W1004 04:23:15.563088 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a44192f_c876_4ef8_835a_ee99845c613e.slice/crio-f5f22edca77a2093cc7d5d42aa82e5ae76c0ec62912a629d39ad492fec30bb93 WatchSource:0}: Error finding container f5f22edca77a2093cc7d5d42aa82e5ae76c0ec62912a629d39ad492fec30bb93: Status 404 returned error can't find the container with id f5f22edca77a2093cc7d5d42aa82e5ae76c0ec62912a629d39ad492fec30bb93 Oct 04 04:23:15 crc kubenswrapper[4770]: I1004 04:23:15.608838 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"5a44192f-c876-4ef8-835a-ee99845c613e","Type":"ContainerStarted","Data":"f5f22edca77a2093cc7d5d42aa82e5ae76c0ec62912a629d39ad492fec30bb93"} Oct 04 04:23:15 crc kubenswrapper[4770]: I1004 04:23:15.699237 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5" path="/var/lib/kubelet/pods/957ae570-7ea4-4fb4-a73f-4ca6b8ae08b5/volumes" Oct 04 04:23:16 crc kubenswrapper[4770]: I1004 04:23:16.621233 4770 generic.go:334] "Generic (PLEG): container finished" podID="5a44192f-c876-4ef8-835a-ee99845c613e" containerID="232e4c7cfbfec43df63d8b53eba51f3c0fefad193e85529bc010f2ff6183b3c5" exitCode=0 Oct 04 04:23:16 crc kubenswrapper[4770]: I1004 04:23:16.621274 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"5a44192f-c876-4ef8-835a-ee99845c613e","Type":"ContainerDied","Data":"232e4c7cfbfec43df63d8b53eba51f3c0fefad193e85529bc010f2ff6183b3c5"} Oct 04 04:23:18 crc kubenswrapper[4770]: I1004 04:23:18.027512 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 04 04:23:18 crc kubenswrapper[4770]: I1004 04:23:18.048050 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_5a44192f-c876-4ef8-835a-ee99845c613e/mariadb-client-4-default/0.log" Oct 04 04:23:18 crc kubenswrapper[4770]: I1004 04:23:18.076952 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 04 04:23:18 crc kubenswrapper[4770]: I1004 04:23:18.081871 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 04 04:23:18 crc kubenswrapper[4770]: I1004 04:23:18.104211 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mz5g\" (UniqueName: \"kubernetes.io/projected/5a44192f-c876-4ef8-835a-ee99845c613e-kube-api-access-2mz5g\") pod \"5a44192f-c876-4ef8-835a-ee99845c613e\" (UID: \"5a44192f-c876-4ef8-835a-ee99845c613e\") " Oct 04 04:23:18 crc kubenswrapper[4770]: I1004 04:23:18.109621 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a44192f-c876-4ef8-835a-ee99845c613e-kube-api-access-2mz5g" (OuterVolumeSpecName: "kube-api-access-2mz5g") pod "5a44192f-c876-4ef8-835a-ee99845c613e" (UID: "5a44192f-c876-4ef8-835a-ee99845c613e"). InnerVolumeSpecName "kube-api-access-2mz5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:23:18 crc kubenswrapper[4770]: I1004 04:23:18.206622 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mz5g\" (UniqueName: \"kubernetes.io/projected/5a44192f-c876-4ef8-835a-ee99845c613e-kube-api-access-2mz5g\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:18 crc kubenswrapper[4770]: I1004 04:23:18.636818 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5f22edca77a2093cc7d5d42aa82e5ae76c0ec62912a629d39ad492fec30bb93" Oct 04 04:23:18 crc kubenswrapper[4770]: I1004 04:23:18.636855 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 04 04:23:19 crc kubenswrapper[4770]: I1004 04:23:19.679698 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:23:19 crc kubenswrapper[4770]: E1004 04:23:19.680256 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:23:19 crc kubenswrapper[4770]: I1004 04:23:19.691475 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a44192f-c876-4ef8-835a-ee99845c613e" path="/var/lib/kubelet/pods/5a44192f-c876-4ef8-835a-ee99845c613e/volumes" Oct 04 04:23:22 crc kubenswrapper[4770]: I1004 04:23:22.689608 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 04 04:23:22 crc kubenswrapper[4770]: E1004 04:23:22.690682 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a44192f-c876-4ef8-835a-ee99845c613e" containerName="mariadb-client-4-default" Oct 04 04:23:22 crc kubenswrapper[4770]: I1004 04:23:22.690702 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a44192f-c876-4ef8-835a-ee99845c613e" containerName="mariadb-client-4-default" Oct 04 04:23:22 crc kubenswrapper[4770]: I1004 04:23:22.690932 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a44192f-c876-4ef8-835a-ee99845c613e" containerName="mariadb-client-4-default" Oct 04 04:23:22 crc kubenswrapper[4770]: I1004 04:23:22.692196 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 04 04:23:22 crc kubenswrapper[4770]: I1004 04:23:22.698230 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t7x2l" Oct 04 04:23:22 crc kubenswrapper[4770]: I1004 04:23:22.711208 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 04 04:23:22 crc kubenswrapper[4770]: I1004 04:23:22.802068 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wnhd\" (UniqueName: \"kubernetes.io/projected/eaf3ad4d-8911-471a-b7ec-eb983225453e-kube-api-access-8wnhd\") pod \"mariadb-client-5-default\" (UID: \"eaf3ad4d-8911-471a-b7ec-eb983225453e\") " pod="openstack/mariadb-client-5-default" Oct 04 04:23:22 crc kubenswrapper[4770]: I1004 04:23:22.903463 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wnhd\" (UniqueName: \"kubernetes.io/projected/eaf3ad4d-8911-471a-b7ec-eb983225453e-kube-api-access-8wnhd\") pod \"mariadb-client-5-default\" (UID: \"eaf3ad4d-8911-471a-b7ec-eb983225453e\") " pod="openstack/mariadb-client-5-default" Oct 04 04:23:22 crc kubenswrapper[4770]: I1004 04:23:22.922928 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wnhd\" (UniqueName: \"kubernetes.io/projected/eaf3ad4d-8911-471a-b7ec-eb983225453e-kube-api-access-8wnhd\") pod \"mariadb-client-5-default\" (UID: \"eaf3ad4d-8911-471a-b7ec-eb983225453e\") " pod="openstack/mariadb-client-5-default" Oct 04 04:23:23 crc kubenswrapper[4770]: I1004 04:23:23.015575 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 04 04:23:23 crc kubenswrapper[4770]: I1004 04:23:23.405497 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 04 04:23:23 crc kubenswrapper[4770]: W1004 04:23:23.412400 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeaf3ad4d_8911_471a_b7ec_eb983225453e.slice/crio-9a302f10d9c9bea8c1b2d4bf8e20f5c48aef6385d383d8dd07d4033ec69a7f75 WatchSource:0}: Error finding container 9a302f10d9c9bea8c1b2d4bf8e20f5c48aef6385d383d8dd07d4033ec69a7f75: Status 404 returned error can't find the container with id 9a302f10d9c9bea8c1b2d4bf8e20f5c48aef6385d383d8dd07d4033ec69a7f75 Oct 04 04:23:23 crc kubenswrapper[4770]: I1004 04:23:23.689630 4770 generic.go:334] "Generic (PLEG): container finished" podID="eaf3ad4d-8911-471a-b7ec-eb983225453e" containerID="8146012b9c65a78b65d9bc6ce7bb69fc3d464136d086056f02310dc3d52a3742" exitCode=0 Oct 04 04:23:23 crc kubenswrapper[4770]: I1004 04:23:23.689687 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"eaf3ad4d-8911-471a-b7ec-eb983225453e","Type":"ContainerDied","Data":"8146012b9c65a78b65d9bc6ce7bb69fc3d464136d086056f02310dc3d52a3742"} Oct 04 04:23:23 crc kubenswrapper[4770]: I1004 04:23:23.689721 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"eaf3ad4d-8911-471a-b7ec-eb983225453e","Type":"ContainerStarted","Data":"9a302f10d9c9bea8c1b2d4bf8e20f5c48aef6385d383d8dd07d4033ec69a7f75"} Oct 04 04:23:24 crc kubenswrapper[4770]: I1004 04:23:24.308898 4770 scope.go:117] "RemoveContainer" containerID="4c9a5976fc39b490147d3bed73c1d8405c85bd5695e2dfa53b7d95d4222e996b" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.116512 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.135933 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wnhd\" (UniqueName: \"kubernetes.io/projected/eaf3ad4d-8911-471a-b7ec-eb983225453e-kube-api-access-8wnhd\") pod \"eaf3ad4d-8911-471a-b7ec-eb983225453e\" (UID: \"eaf3ad4d-8911-471a-b7ec-eb983225453e\") " Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.136299 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_eaf3ad4d-8911-471a-b7ec-eb983225453e/mariadb-client-5-default/0.log" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.144305 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaf3ad4d-8911-471a-b7ec-eb983225453e-kube-api-access-8wnhd" (OuterVolumeSpecName: "kube-api-access-8wnhd") pod "eaf3ad4d-8911-471a-b7ec-eb983225453e" (UID: "eaf3ad4d-8911-471a-b7ec-eb983225453e"). InnerVolumeSpecName "kube-api-access-8wnhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.172207 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.180079 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.239056 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wnhd\" (UniqueName: \"kubernetes.io/projected/eaf3ad4d-8911-471a-b7ec-eb983225453e-kube-api-access-8wnhd\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.382158 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 04 04:23:25 crc kubenswrapper[4770]: E1004 04:23:25.382713 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf3ad4d-8911-471a-b7ec-eb983225453e" containerName="mariadb-client-5-default" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.382749 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf3ad4d-8911-471a-b7ec-eb983225453e" containerName="mariadb-client-5-default" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.383061 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf3ad4d-8911-471a-b7ec-eb983225453e" containerName="mariadb-client-5-default" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.384126 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.409191 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.442831 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2skzd\" (UniqueName: \"kubernetes.io/projected/48377739-d52d-454b-a47f-6f3c1b09f7ef-kube-api-access-2skzd\") pod \"mariadb-client-6-default\" (UID: \"48377739-d52d-454b-a47f-6f3c1b09f7ef\") " pod="openstack/mariadb-client-6-default" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.545025 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2skzd\" (UniqueName: \"kubernetes.io/projected/48377739-d52d-454b-a47f-6f3c1b09f7ef-kube-api-access-2skzd\") pod \"mariadb-client-6-default\" (UID: \"48377739-d52d-454b-a47f-6f3c1b09f7ef\") " pod="openstack/mariadb-client-6-default" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.588514 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2skzd\" (UniqueName: \"kubernetes.io/projected/48377739-d52d-454b-a47f-6f3c1b09f7ef-kube-api-access-2skzd\") pod \"mariadb-client-6-default\" (UID: \"48377739-d52d-454b-a47f-6f3c1b09f7ef\") " pod="openstack/mariadb-client-6-default" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.685679 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaf3ad4d-8911-471a-b7ec-eb983225453e" path="/var/lib/kubelet/pods/eaf3ad4d-8911-471a-b7ec-eb983225453e/volumes" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.708289 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.712353 4770 scope.go:117] "RemoveContainer" containerID="8146012b9c65a78b65d9bc6ce7bb69fc3d464136d086056f02310dc3d52a3742" Oct 04 04:23:25 crc kubenswrapper[4770]: I1004 04:23:25.712554 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 04 04:23:26 crc kubenswrapper[4770]: I1004 04:23:26.237699 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 04 04:23:26 crc kubenswrapper[4770]: W1004 04:23:26.239181 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48377739_d52d_454b_a47f_6f3c1b09f7ef.slice/crio-ba7469bfe39e74705a7725c1b30653fe433bf3e893070df7bc8e29a113254122 WatchSource:0}: Error finding container ba7469bfe39e74705a7725c1b30653fe433bf3e893070df7bc8e29a113254122: Status 404 returned error can't find the container with id ba7469bfe39e74705a7725c1b30653fe433bf3e893070df7bc8e29a113254122 Oct 04 04:23:26 crc kubenswrapper[4770]: I1004 04:23:26.723850 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"48377739-d52d-454b-a47f-6f3c1b09f7ef","Type":"ContainerStarted","Data":"b6dd244a3a3746399b9b853f22a90e23bde1b8405e091278bc0d2ba2543b14e3"} Oct 04 04:23:26 crc kubenswrapper[4770]: I1004 04:23:26.723923 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"48377739-d52d-454b-a47f-6f3c1b09f7ef","Type":"ContainerStarted","Data":"ba7469bfe39e74705a7725c1b30653fe433bf3e893070df7bc8e29a113254122"} Oct 04 04:23:27 crc kubenswrapper[4770]: I1004 04:23:27.736673 4770 generic.go:334] "Generic (PLEG): container finished" podID="48377739-d52d-454b-a47f-6f3c1b09f7ef" containerID="b6dd244a3a3746399b9b853f22a90e23bde1b8405e091278bc0d2ba2543b14e3" exitCode=0 Oct 04 04:23:27 crc kubenswrapper[4770]: I1004 04:23:27.736781 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"48377739-d52d-454b-a47f-6f3c1b09f7ef","Type":"ContainerDied","Data":"b6dd244a3a3746399b9b853f22a90e23bde1b8405e091278bc0d2ba2543b14e3"} Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.212708 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.279813 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_48377739-d52d-454b-a47f-6f3c1b09f7ef/mariadb-client-6-default/0.log" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.308225 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.314532 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.324726 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2skzd\" (UniqueName: \"kubernetes.io/projected/48377739-d52d-454b-a47f-6f3c1b09f7ef-kube-api-access-2skzd\") pod \"48377739-d52d-454b-a47f-6f3c1b09f7ef\" (UID: \"48377739-d52d-454b-a47f-6f3c1b09f7ef\") " Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.331241 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48377739-d52d-454b-a47f-6f3c1b09f7ef-kube-api-access-2skzd" (OuterVolumeSpecName: "kube-api-access-2skzd") pod "48377739-d52d-454b-a47f-6f3c1b09f7ef" (UID: "48377739-d52d-454b-a47f-6f3c1b09f7ef"). InnerVolumeSpecName "kube-api-access-2skzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.427150 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2skzd\" (UniqueName: \"kubernetes.io/projected/48377739-d52d-454b-a47f-6f3c1b09f7ef-kube-api-access-2skzd\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.529091 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 04 04:23:29 crc kubenswrapper[4770]: E1004 04:23:29.529455 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48377739-d52d-454b-a47f-6f3c1b09f7ef" containerName="mariadb-client-6-default" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.529474 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="48377739-d52d-454b-a47f-6f3c1b09f7ef" containerName="mariadb-client-6-default" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.529619 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="48377739-d52d-454b-a47f-6f3c1b09f7ef" containerName="mariadb-client-6-default" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.530134 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.534296 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.630750 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6xts\" (UniqueName: \"kubernetes.io/projected/c025592e-2af0-4a99-8dab-16585886c133-kube-api-access-s6xts\") pod \"mariadb-client-7-default\" (UID: \"c025592e-2af0-4a99-8dab-16585886c133\") " pod="openstack/mariadb-client-7-default" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.689734 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48377739-d52d-454b-a47f-6f3c1b09f7ef" path="/var/lib/kubelet/pods/48377739-d52d-454b-a47f-6f3c1b09f7ef/volumes" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.732837 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6xts\" (UniqueName: \"kubernetes.io/projected/c025592e-2af0-4a99-8dab-16585886c133-kube-api-access-s6xts\") pod \"mariadb-client-7-default\" (UID: \"c025592e-2af0-4a99-8dab-16585886c133\") " pod="openstack/mariadb-client-7-default" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.751981 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6xts\" (UniqueName: \"kubernetes.io/projected/c025592e-2af0-4a99-8dab-16585886c133-kube-api-access-s6xts\") pod \"mariadb-client-7-default\" (UID: \"c025592e-2af0-4a99-8dab-16585886c133\") " pod="openstack/mariadb-client-7-default" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.765509 4770 scope.go:117] "RemoveContainer" containerID="b6dd244a3a3746399b9b853f22a90e23bde1b8405e091278bc0d2ba2543b14e3" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.765681 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 04 04:23:29 crc kubenswrapper[4770]: I1004 04:23:29.860130 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 04 04:23:30 crc kubenswrapper[4770]: I1004 04:23:30.406246 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 04 04:23:30 crc kubenswrapper[4770]: W1004 04:23:30.457907 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc025592e_2af0_4a99_8dab_16585886c133.slice/crio-17d8f7619d6ab6c352003d4d1b5ec612b5d8bd32c51301949ba3ecd371fd7e2c WatchSource:0}: Error finding container 17d8f7619d6ab6c352003d4d1b5ec612b5d8bd32c51301949ba3ecd371fd7e2c: Status 404 returned error can't find the container with id 17d8f7619d6ab6c352003d4d1b5ec612b5d8bd32c51301949ba3ecd371fd7e2c Oct 04 04:23:30 crc kubenswrapper[4770]: I1004 04:23:30.777653 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"c025592e-2af0-4a99-8dab-16585886c133","Type":"ContainerStarted","Data":"d6a1a72915e60df6888ce9c2681434043035e8b3aed085df8710d9d7ef0e62ca"} Oct 04 04:23:30 crc kubenswrapper[4770]: I1004 04:23:30.777750 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"c025592e-2af0-4a99-8dab-16585886c133","Type":"ContainerStarted","Data":"17d8f7619d6ab6c352003d4d1b5ec612b5d8bd32c51301949ba3ecd371fd7e2c"} Oct 04 04:23:31 crc kubenswrapper[4770]: I1004 04:23:31.794279 4770 generic.go:334] "Generic (PLEG): container finished" podID="c025592e-2af0-4a99-8dab-16585886c133" containerID="d6a1a72915e60df6888ce9c2681434043035e8b3aed085df8710d9d7ef0e62ca" exitCode=0 Oct 04 04:23:31 crc kubenswrapper[4770]: I1004 04:23:31.794344 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"c025592e-2af0-4a99-8dab-16585886c133","Type":"ContainerDied","Data":"d6a1a72915e60df6888ce9c2681434043035e8b3aed085df8710d9d7ef0e62ca"} Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.198531 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.220665 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_c025592e-2af0-4a99-8dab-16585886c133/mariadb-client-7-default/0.log" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.250414 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.256863 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.289155 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6xts\" (UniqueName: \"kubernetes.io/projected/c025592e-2af0-4a99-8dab-16585886c133-kube-api-access-s6xts\") pod \"c025592e-2af0-4a99-8dab-16585886c133\" (UID: \"c025592e-2af0-4a99-8dab-16585886c133\") " Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.295669 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c025592e-2af0-4a99-8dab-16585886c133-kube-api-access-s6xts" (OuterVolumeSpecName: "kube-api-access-s6xts") pod "c025592e-2af0-4a99-8dab-16585886c133" (UID: "c025592e-2af0-4a99-8dab-16585886c133"). InnerVolumeSpecName "kube-api-access-s6xts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.391487 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6xts\" (UniqueName: \"kubernetes.io/projected/c025592e-2af0-4a99-8dab-16585886c133-kube-api-access-s6xts\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.463132 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 04 04:23:33 crc kubenswrapper[4770]: E1004 04:23:33.464995 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c025592e-2af0-4a99-8dab-16585886c133" containerName="mariadb-client-7-default" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.465087 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c025592e-2af0-4a99-8dab-16585886c133" containerName="mariadb-client-7-default" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.465524 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c025592e-2af0-4a99-8dab-16585886c133" containerName="mariadb-client-7-default" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.466901 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.471002 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.594927 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hd2r\" (UniqueName: \"kubernetes.io/projected/e042c14c-7ec3-4e81-a0dc-bea53c1be6f1-kube-api-access-2hd2r\") pod \"mariadb-client-2\" (UID: \"e042c14c-7ec3-4e81-a0dc-bea53c1be6f1\") " pod="openstack/mariadb-client-2" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.673451 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:23:33 crc kubenswrapper[4770]: E1004 04:23:33.674068 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.687791 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c025592e-2af0-4a99-8dab-16585886c133" path="/var/lib/kubelet/pods/c025592e-2af0-4a99-8dab-16585886c133/volumes" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.696960 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hd2r\" (UniqueName: \"kubernetes.io/projected/e042c14c-7ec3-4e81-a0dc-bea53c1be6f1-kube-api-access-2hd2r\") pod \"mariadb-client-2\" (UID: \"e042c14c-7ec3-4e81-a0dc-bea53c1be6f1\") " pod="openstack/mariadb-client-2" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.715604 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hd2r\" (UniqueName: \"kubernetes.io/projected/e042c14c-7ec3-4e81-a0dc-bea53c1be6f1-kube-api-access-2hd2r\") pod \"mariadb-client-2\" (UID: \"e042c14c-7ec3-4e81-a0dc-bea53c1be6f1\") " pod="openstack/mariadb-client-2" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.792652 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.829818 4770 scope.go:117] "RemoveContainer" containerID="d6a1a72915e60df6888ce9c2681434043035e8b3aed085df8710d9d7ef0e62ca" Oct 04 04:23:33 crc kubenswrapper[4770]: I1004 04:23:33.829945 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 04 04:23:34 crc kubenswrapper[4770]: I1004 04:23:34.396146 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 04 04:23:34 crc kubenswrapper[4770]: W1004 04:23:34.399333 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode042c14c_7ec3_4e81_a0dc_bea53c1be6f1.slice/crio-0d04cf2d48672a8a84eec771cf40dbaa50d0248783b0ea13e234e3d806e4702a WatchSource:0}: Error finding container 0d04cf2d48672a8a84eec771cf40dbaa50d0248783b0ea13e234e3d806e4702a: Status 404 returned error can't find the container with id 0d04cf2d48672a8a84eec771cf40dbaa50d0248783b0ea13e234e3d806e4702a Oct 04 04:23:34 crc kubenswrapper[4770]: I1004 04:23:34.842227 4770 generic.go:334] "Generic (PLEG): container finished" podID="e042c14c-7ec3-4e81-a0dc-bea53c1be6f1" containerID="108db819f7b370469975ad54fd596e8fbd090999a05ccdde7714e0f7b2026b04" exitCode=0 Oct 04 04:23:34 crc kubenswrapper[4770]: I1004 04:23:34.842312 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"e042c14c-7ec3-4e81-a0dc-bea53c1be6f1","Type":"ContainerDied","Data":"108db819f7b370469975ad54fd596e8fbd090999a05ccdde7714e0f7b2026b04"} Oct 04 04:23:34 crc kubenswrapper[4770]: I1004 04:23:34.842808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"e042c14c-7ec3-4e81-a0dc-bea53c1be6f1","Type":"ContainerStarted","Data":"0d04cf2d48672a8a84eec771cf40dbaa50d0248783b0ea13e234e3d806e4702a"} Oct 04 04:23:36 crc kubenswrapper[4770]: I1004 04:23:36.249980 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 04 04:23:36 crc kubenswrapper[4770]: I1004 04:23:36.274182 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_e042c14c-7ec3-4e81-a0dc-bea53c1be6f1/mariadb-client-2/0.log" Oct 04 04:23:36 crc kubenswrapper[4770]: I1004 04:23:36.307000 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 04 04:23:36 crc kubenswrapper[4770]: I1004 04:23:36.322042 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 04 04:23:36 crc kubenswrapper[4770]: I1004 04:23:36.343806 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hd2r\" (UniqueName: \"kubernetes.io/projected/e042c14c-7ec3-4e81-a0dc-bea53c1be6f1-kube-api-access-2hd2r\") pod \"e042c14c-7ec3-4e81-a0dc-bea53c1be6f1\" (UID: \"e042c14c-7ec3-4e81-a0dc-bea53c1be6f1\") " Oct 04 04:23:36 crc kubenswrapper[4770]: I1004 04:23:36.351040 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e042c14c-7ec3-4e81-a0dc-bea53c1be6f1-kube-api-access-2hd2r" (OuterVolumeSpecName: "kube-api-access-2hd2r") pod "e042c14c-7ec3-4e81-a0dc-bea53c1be6f1" (UID: "e042c14c-7ec3-4e81-a0dc-bea53c1be6f1"). InnerVolumeSpecName "kube-api-access-2hd2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:23:36 crc kubenswrapper[4770]: I1004 04:23:36.445765 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hd2r\" (UniqueName: \"kubernetes.io/projected/e042c14c-7ec3-4e81-a0dc-bea53c1be6f1-kube-api-access-2hd2r\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:36 crc kubenswrapper[4770]: I1004 04:23:36.867068 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d04cf2d48672a8a84eec771cf40dbaa50d0248783b0ea13e234e3d806e4702a" Oct 04 04:23:36 crc kubenswrapper[4770]: I1004 04:23:36.867091 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 04 04:23:37 crc kubenswrapper[4770]: I1004 04:23:37.684756 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e042c14c-7ec3-4e81-a0dc-bea53c1be6f1" path="/var/lib/kubelet/pods/e042c14c-7ec3-4e81-a0dc-bea53c1be6f1/volumes" Oct 04 04:23:47 crc kubenswrapper[4770]: I1004 04:23:47.677656 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:23:47 crc kubenswrapper[4770]: E1004 04:23:47.679351 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:23:58 crc kubenswrapper[4770]: I1004 04:23:58.675143 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:23:58 crc kubenswrapper[4770]: E1004 04:23:58.676343 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:24:11 crc kubenswrapper[4770]: I1004 04:24:11.673986 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:24:11 crc kubenswrapper[4770]: E1004 04:24:11.674939 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:24:23 crc kubenswrapper[4770]: I1004 04:24:23.673719 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:24:23 crc kubenswrapper[4770]: E1004 04:24:23.674530 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:24:36 crc kubenswrapper[4770]: I1004 04:24:36.682531 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:24:37 crc kubenswrapper[4770]: I1004 04:24:37.494277 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"6ee9dd0aa38f827331a89fde2cd36be108fb6bc2a67c9f1e927459a0a521b3c6"} Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.339460 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-srpng"] Oct 04 04:26:55 crc kubenswrapper[4770]: E1004 04:26:55.357361 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e042c14c-7ec3-4e81-a0dc-bea53c1be6f1" containerName="mariadb-client-2" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.357408 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e042c14c-7ec3-4e81-a0dc-bea53c1be6f1" containerName="mariadb-client-2" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.357927 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e042c14c-7ec3-4e81-a0dc-bea53c1be6f1" containerName="mariadb-client-2" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.362940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.378994 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-srpng"] Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.459557 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bvbv\" (UniqueName: \"kubernetes.io/projected/2436562a-3cf9-45df-862a-b9a969675e2d-kube-api-access-4bvbv\") pod \"redhat-marketplace-srpng\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.459685 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-utilities\") pod \"redhat-marketplace-srpng\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.459725 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-catalog-content\") pod \"redhat-marketplace-srpng\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.560916 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bvbv\" (UniqueName: \"kubernetes.io/projected/2436562a-3cf9-45df-862a-b9a969675e2d-kube-api-access-4bvbv\") pod \"redhat-marketplace-srpng\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.561027 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-utilities\") pod \"redhat-marketplace-srpng\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.561056 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-catalog-content\") pod \"redhat-marketplace-srpng\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.561583 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-catalog-content\") pod \"redhat-marketplace-srpng\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.561706 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-utilities\") pod \"redhat-marketplace-srpng\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.585217 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bvbv\" (UniqueName: \"kubernetes.io/projected/2436562a-3cf9-45df-862a-b9a969675e2d-kube-api-access-4bvbv\") pod \"redhat-marketplace-srpng\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:55 crc kubenswrapper[4770]: I1004 04:26:55.692847 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:26:56 crc kubenswrapper[4770]: I1004 04:26:56.160149 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-srpng"] Oct 04 04:26:56 crc kubenswrapper[4770]: I1004 04:26:56.864495 4770 generic.go:334] "Generic (PLEG): container finished" podID="2436562a-3cf9-45df-862a-b9a969675e2d" containerID="c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f" exitCode=0 Oct 04 04:26:56 crc kubenswrapper[4770]: I1004 04:26:56.864763 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srpng" event={"ID":"2436562a-3cf9-45df-862a-b9a969675e2d","Type":"ContainerDied","Data":"c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f"} Oct 04 04:26:56 crc kubenswrapper[4770]: I1004 04:26:56.864803 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srpng" event={"ID":"2436562a-3cf9-45df-862a-b9a969675e2d","Type":"ContainerStarted","Data":"c6ddd74788374a5d38374118ccc3758ad1e88aee66618fa86f90f524a76f7b22"} Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.149423 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p2lnx"] Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.150988 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.168201 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p2lnx"] Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.287476 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhvdl\" (UniqueName: \"kubernetes.io/projected/6db4349c-ad5c-4141-a219-d1111575ab6a-kube-api-access-bhvdl\") pod \"certified-operators-p2lnx\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.287909 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-utilities\") pod \"certified-operators-p2lnx\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.288063 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-catalog-content\") pod \"certified-operators-p2lnx\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.390132 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-utilities\") pod \"certified-operators-p2lnx\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.390202 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-catalog-content\") pod \"certified-operators-p2lnx\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.390290 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhvdl\" (UniqueName: \"kubernetes.io/projected/6db4349c-ad5c-4141-a219-d1111575ab6a-kube-api-access-bhvdl\") pod \"certified-operators-p2lnx\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.390781 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-utilities\") pod \"certified-operators-p2lnx\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.390897 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-catalog-content\") pod \"certified-operators-p2lnx\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.428280 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhvdl\" (UniqueName: \"kubernetes.io/projected/6db4349c-ad5c-4141-a219-d1111575ab6a-kube-api-access-bhvdl\") pod \"certified-operators-p2lnx\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.510878 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.735101 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-shhkz"] Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.736520 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.750866 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-shhkz"] Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.896262 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bprbx\" (UniqueName: \"kubernetes.io/projected/c27542e5-b249-430d-8360-5453ccbe20c8-kube-api-access-bprbx\") pod \"community-operators-shhkz\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.897484 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-utilities\") pod \"community-operators-shhkz\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.897617 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-catalog-content\") pod \"community-operators-shhkz\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.977864 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p2lnx"] Oct 04 04:26:57 crc kubenswrapper[4770]: W1004 04:26:57.984496 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6db4349c_ad5c_4141_a219_d1111575ab6a.slice/crio-7e853866381fbdec69fe4fa65dec3e0ebcca289a334e3248dde6108d6cb5ace7 WatchSource:0}: Error finding container 7e853866381fbdec69fe4fa65dec3e0ebcca289a334e3248dde6108d6cb5ace7: Status 404 returned error can't find the container with id 7e853866381fbdec69fe4fa65dec3e0ebcca289a334e3248dde6108d6cb5ace7 Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.999158 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-utilities\") pod \"community-operators-shhkz\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.999227 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-catalog-content\") pod \"community-operators-shhkz\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:57 crc kubenswrapper[4770]: I1004 04:26:57.999313 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bprbx\" (UniqueName: \"kubernetes.io/projected/c27542e5-b249-430d-8360-5453ccbe20c8-kube-api-access-bprbx\") pod \"community-operators-shhkz\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:57.999997 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-utilities\") pod \"community-operators-shhkz\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.000158 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-catalog-content\") pod \"community-operators-shhkz\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.020997 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bprbx\" (UniqueName: \"kubernetes.io/projected/c27542e5-b249-430d-8360-5453ccbe20c8-kube-api-access-bprbx\") pod \"community-operators-shhkz\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.056699 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.359472 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-shhkz"] Oct 04 04:26:58 crc kubenswrapper[4770]: W1004 04:26:58.447491 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc27542e5_b249_430d_8360_5453ccbe20c8.slice/crio-7e6cf514b2c55ad4c1721256e738361b4159e2642697d17e815c0678dfb45aaa WatchSource:0}: Error finding container 7e6cf514b2c55ad4c1721256e738361b4159e2642697d17e815c0678dfb45aaa: Status 404 returned error can't find the container with id 7e6cf514b2c55ad4c1721256e738361b4159e2642697d17e815c0678dfb45aaa Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.891117 4770 generic.go:334] "Generic (PLEG): container finished" podID="c27542e5-b249-430d-8360-5453ccbe20c8" containerID="c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a" exitCode=0 Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.891335 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-shhkz" event={"ID":"c27542e5-b249-430d-8360-5453ccbe20c8","Type":"ContainerDied","Data":"c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a"} Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.891423 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-shhkz" event={"ID":"c27542e5-b249-430d-8360-5453ccbe20c8","Type":"ContainerStarted","Data":"7e6cf514b2c55ad4c1721256e738361b4159e2642697d17e815c0678dfb45aaa"} Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.894175 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2lnx" event={"ID":"6db4349c-ad5c-4141-a219-d1111575ab6a","Type":"ContainerDied","Data":"5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131"} Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.894139 4770 generic.go:334] "Generic (PLEG): container finished" podID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerID="5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131" exitCode=0 Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.894692 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2lnx" event={"ID":"6db4349c-ad5c-4141-a219-d1111575ab6a","Type":"ContainerStarted","Data":"7e853866381fbdec69fe4fa65dec3e0ebcca289a334e3248dde6108d6cb5ace7"} Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.898614 4770 generic.go:334] "Generic (PLEG): container finished" podID="2436562a-3cf9-45df-862a-b9a969675e2d" containerID="11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034" exitCode=0 Oct 04 04:26:58 crc kubenswrapper[4770]: I1004 04:26:58.898658 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srpng" event={"ID":"2436562a-3cf9-45df-862a-b9a969675e2d","Type":"ContainerDied","Data":"11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034"} Oct 04 04:26:59 crc kubenswrapper[4770]: I1004 04:26:59.906930 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srpng" event={"ID":"2436562a-3cf9-45df-862a-b9a969675e2d","Type":"ContainerStarted","Data":"a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513"} Oct 04 04:27:00 crc kubenswrapper[4770]: I1004 04:27:00.917512 4770 generic.go:334] "Generic (PLEG): container finished" podID="c27542e5-b249-430d-8360-5453ccbe20c8" containerID="75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a" exitCode=0 Oct 04 04:27:00 crc kubenswrapper[4770]: I1004 04:27:00.917941 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-shhkz" event={"ID":"c27542e5-b249-430d-8360-5453ccbe20c8","Type":"ContainerDied","Data":"75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a"} Oct 04 04:27:00 crc kubenswrapper[4770]: I1004 04:27:00.924795 4770 generic.go:334] "Generic (PLEG): container finished" podID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerID="1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e" exitCode=0 Oct 04 04:27:00 crc kubenswrapper[4770]: I1004 04:27:00.924857 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2lnx" event={"ID":"6db4349c-ad5c-4141-a219-d1111575ab6a","Type":"ContainerDied","Data":"1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e"} Oct 04 04:27:00 crc kubenswrapper[4770]: I1004 04:27:00.954812 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-srpng" podStartSLOduration=3.395522123 podStartE2EDuration="5.954779262s" podCreationTimestamp="2025-10-04 04:26:55 +0000 UTC" firstStartedPulling="2025-10-04 04:26:56.866234914 +0000 UTC m=+5028.158244636" lastFinishedPulling="2025-10-04 04:26:59.425492063 +0000 UTC m=+5030.717501775" observedRunningTime="2025-10-04 04:26:59.92774847 +0000 UTC m=+5031.219758192" watchObservedRunningTime="2025-10-04 04:27:00.954779262 +0000 UTC m=+5032.246789014" Oct 04 04:27:01 crc kubenswrapper[4770]: I1004 04:27:01.795648 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:27:01 crc kubenswrapper[4770]: I1004 04:27:01.796084 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:27:01 crc kubenswrapper[4770]: I1004 04:27:01.937633 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-shhkz" event={"ID":"c27542e5-b249-430d-8360-5453ccbe20c8","Type":"ContainerStarted","Data":"645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29"} Oct 04 04:27:01 crc kubenswrapper[4770]: I1004 04:27:01.940983 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2lnx" event={"ID":"6db4349c-ad5c-4141-a219-d1111575ab6a","Type":"ContainerStarted","Data":"f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb"} Oct 04 04:27:01 crc kubenswrapper[4770]: I1004 04:27:01.995735 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-shhkz" podStartSLOduration=2.5746389499999998 podStartE2EDuration="4.995704428s" podCreationTimestamp="2025-10-04 04:26:57 +0000 UTC" firstStartedPulling="2025-10-04 04:26:58.902484733 +0000 UTC m=+5030.194494465" lastFinishedPulling="2025-10-04 04:27:01.323550171 +0000 UTC m=+5032.615559943" observedRunningTime="2025-10-04 04:27:01.962845879 +0000 UTC m=+5033.254855621" watchObservedRunningTime="2025-10-04 04:27:01.995704428 +0000 UTC m=+5033.287714150" Oct 04 04:27:01 crc kubenswrapper[4770]: I1004 04:27:01.997565 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p2lnx" podStartSLOduration=2.435931566 podStartE2EDuration="4.997553457s" podCreationTimestamp="2025-10-04 04:26:57 +0000 UTC" firstStartedPulling="2025-10-04 04:26:58.902457693 +0000 UTC m=+5030.194467405" lastFinishedPulling="2025-10-04 04:27:01.464079554 +0000 UTC m=+5032.756089296" observedRunningTime="2025-10-04 04:27:01.989394383 +0000 UTC m=+5033.281404115" watchObservedRunningTime="2025-10-04 04:27:01.997553457 +0000 UTC m=+5033.289563189" Oct 04 04:27:05 crc kubenswrapper[4770]: I1004 04:27:05.693098 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:27:05 crc kubenswrapper[4770]: I1004 04:27:05.693666 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:27:05 crc kubenswrapper[4770]: I1004 04:27:05.778614 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:27:06 crc kubenswrapper[4770]: I1004 04:27:06.041554 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:27:07 crc kubenswrapper[4770]: I1004 04:27:07.130348 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-srpng"] Oct 04 04:27:07 crc kubenswrapper[4770]: I1004 04:27:07.511644 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:27:07 crc kubenswrapper[4770]: I1004 04:27:07.511980 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:27:07 crc kubenswrapper[4770]: I1004 04:27:07.595209 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:27:08 crc kubenswrapper[4770]: I1004 04:27:08.000730 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-srpng" podUID="2436562a-3cf9-45df-862a-b9a969675e2d" containerName="registry-server" containerID="cri-o://a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513" gracePeriod=2 Oct 04 04:27:08 crc kubenswrapper[4770]: I1004 04:27:08.057054 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:27:08 crc kubenswrapper[4770]: I1004 04:27:08.057118 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:27:08 crc kubenswrapper[4770]: I1004 04:27:08.061350 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:27:08 crc kubenswrapper[4770]: I1004 04:27:08.147713 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:27:08 crc kubenswrapper[4770]: I1004 04:27:08.928594 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.007436 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bvbv\" (UniqueName: \"kubernetes.io/projected/2436562a-3cf9-45df-862a-b9a969675e2d-kube-api-access-4bvbv\") pod \"2436562a-3cf9-45df-862a-b9a969675e2d\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.007575 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-utilities\") pod \"2436562a-3cf9-45df-862a-b9a969675e2d\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.007624 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-catalog-content\") pod \"2436562a-3cf9-45df-862a-b9a969675e2d\" (UID: \"2436562a-3cf9-45df-862a-b9a969675e2d\") " Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.010467 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-utilities" (OuterVolumeSpecName: "utilities") pod "2436562a-3cf9-45df-862a-b9a969675e2d" (UID: "2436562a-3cf9-45df-862a-b9a969675e2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.014118 4770 generic.go:334] "Generic (PLEG): container finished" podID="2436562a-3cf9-45df-862a-b9a969675e2d" containerID="a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513" exitCode=0 Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.014342 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srpng" event={"ID":"2436562a-3cf9-45df-862a-b9a969675e2d","Type":"ContainerDied","Data":"a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513"} Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.014386 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srpng" event={"ID":"2436562a-3cf9-45df-862a-b9a969675e2d","Type":"ContainerDied","Data":"c6ddd74788374a5d38374118ccc3758ad1e88aee66618fa86f90f524a76f7b22"} Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.014405 4770 scope.go:117] "RemoveContainer" containerID="a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.014613 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srpng" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.017780 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2436562a-3cf9-45df-862a-b9a969675e2d-kube-api-access-4bvbv" (OuterVolumeSpecName: "kube-api-access-4bvbv") pod "2436562a-3cf9-45df-862a-b9a969675e2d" (UID: "2436562a-3cf9-45df-862a-b9a969675e2d"). InnerVolumeSpecName "kube-api-access-4bvbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.030821 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2436562a-3cf9-45df-862a-b9a969675e2d" (UID: "2436562a-3cf9-45df-862a-b9a969675e2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.098785 4770 scope.go:117] "RemoveContainer" containerID="11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.109109 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.109154 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2436562a-3cf9-45df-862a-b9a969675e2d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.109170 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bvbv\" (UniqueName: \"kubernetes.io/projected/2436562a-3cf9-45df-862a-b9a969675e2d-kube-api-access-4bvbv\") on node \"crc\" DevicePath \"\"" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.116881 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.131765 4770 scope.go:117] "RemoveContainer" containerID="c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.180830 4770 scope.go:117] "RemoveContainer" containerID="a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513" Oct 04 04:27:09 crc kubenswrapper[4770]: E1004 04:27:09.181661 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513\": container with ID starting with a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513 not found: ID does not exist" containerID="a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.181716 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513"} err="failed to get container status \"a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513\": rpc error: code = NotFound desc = could not find container \"a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513\": container with ID starting with a2d5cf6ba11618da99d204886805eb6919ae3341222301e5e3b880f1dd305513 not found: ID does not exist" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.181747 4770 scope.go:117] "RemoveContainer" containerID="11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034" Oct 04 04:27:09 crc kubenswrapper[4770]: E1004 04:27:09.182276 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034\": container with ID starting with 11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034 not found: ID does not exist" containerID="11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.182324 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034"} err="failed to get container status \"11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034\": rpc error: code = NotFound desc = could not find container \"11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034\": container with ID starting with 11bed82a2798cc7cd273410e6822c2d6f85f5700d21d0c734223c499b8d3c034 not found: ID does not exist" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.182354 4770 scope.go:117] "RemoveContainer" containerID="c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f" Oct 04 04:27:09 crc kubenswrapper[4770]: E1004 04:27:09.182624 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f\": container with ID starting with c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f not found: ID does not exist" containerID="c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.182653 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f"} err="failed to get container status \"c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f\": rpc error: code = NotFound desc = could not find container \"c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f\": container with ID starting with c78f8c7c67086afcc994be19dc4f8c5c60a523dd22b5b9fc3e7980d87cb7389f not found: ID does not exist" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.342874 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-srpng"] Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.349827 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-srpng"] Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.696429 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2436562a-3cf9-45df-862a-b9a969675e2d" path="/var/lib/kubelet/pods/2436562a-3cf9-45df-862a-b9a969675e2d/volumes" Oct 04 04:27:09 crc kubenswrapper[4770]: I1004 04:27:09.935739 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p2lnx"] Oct 04 04:27:10 crc kubenswrapper[4770]: I1004 04:27:10.033984 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p2lnx" podUID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerName="registry-server" containerID="cri-o://f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb" gracePeriod=2 Oct 04 04:27:10 crc kubenswrapper[4770]: I1004 04:27:10.791818 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:27:10 crc kubenswrapper[4770]: I1004 04:27:10.834841 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-utilities\") pod \"6db4349c-ad5c-4141-a219-d1111575ab6a\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " Oct 04 04:27:10 crc kubenswrapper[4770]: I1004 04:27:10.834910 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-catalog-content\") pod \"6db4349c-ad5c-4141-a219-d1111575ab6a\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " Oct 04 04:27:10 crc kubenswrapper[4770]: I1004 04:27:10.835103 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhvdl\" (UniqueName: \"kubernetes.io/projected/6db4349c-ad5c-4141-a219-d1111575ab6a-kube-api-access-bhvdl\") pod \"6db4349c-ad5c-4141-a219-d1111575ab6a\" (UID: \"6db4349c-ad5c-4141-a219-d1111575ab6a\") " Oct 04 04:27:10 crc kubenswrapper[4770]: I1004 04:27:10.837127 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-utilities" (OuterVolumeSpecName: "utilities") pod "6db4349c-ad5c-4141-a219-d1111575ab6a" (UID: "6db4349c-ad5c-4141-a219-d1111575ab6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:27:10 crc kubenswrapper[4770]: I1004 04:27:10.841579 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6db4349c-ad5c-4141-a219-d1111575ab6a-kube-api-access-bhvdl" (OuterVolumeSpecName: "kube-api-access-bhvdl") pod "6db4349c-ad5c-4141-a219-d1111575ab6a" (UID: "6db4349c-ad5c-4141-a219-d1111575ab6a"). InnerVolumeSpecName "kube-api-access-bhvdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:27:10 crc kubenswrapper[4770]: I1004 04:27:10.938236 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:27:10 crc kubenswrapper[4770]: I1004 04:27:10.938275 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhvdl\" (UniqueName: \"kubernetes.io/projected/6db4349c-ad5c-4141-a219-d1111575ab6a-kube-api-access-bhvdl\") on node \"crc\" DevicePath \"\"" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.045989 4770 generic.go:334] "Generic (PLEG): container finished" podID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerID="f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb" exitCode=0 Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.046116 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2lnx" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.046113 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2lnx" event={"ID":"6db4349c-ad5c-4141-a219-d1111575ab6a","Type":"ContainerDied","Data":"f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb"} Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.046248 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2lnx" event={"ID":"6db4349c-ad5c-4141-a219-d1111575ab6a","Type":"ContainerDied","Data":"7e853866381fbdec69fe4fa65dec3e0ebcca289a334e3248dde6108d6cb5ace7"} Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.046267 4770 scope.go:117] "RemoveContainer" containerID="f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.065684 4770 scope.go:117] "RemoveContainer" containerID="1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.084589 4770 scope.go:117] "RemoveContainer" containerID="5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.132088 4770 scope.go:117] "RemoveContainer" containerID="f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb" Oct 04 04:27:11 crc kubenswrapper[4770]: E1004 04:27:11.132653 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb\": container with ID starting with f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb not found: ID does not exist" containerID="f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.132712 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb"} err="failed to get container status \"f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb\": rpc error: code = NotFound desc = could not find container \"f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb\": container with ID starting with f0b4dd2688ab233bc5ea3698c05fb2e394607c773451f55df845d7065e1237eb not found: ID does not exist" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.132746 4770 scope.go:117] "RemoveContainer" containerID="1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e" Oct 04 04:27:11 crc kubenswrapper[4770]: E1004 04:27:11.133165 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e\": container with ID starting with 1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e not found: ID does not exist" containerID="1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.133192 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e"} err="failed to get container status \"1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e\": rpc error: code = NotFound desc = could not find container \"1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e\": container with ID starting with 1b6d88a25de95815b819ae5cd6d009a3a121ec3883bfe2fee09f92208590887e not found: ID does not exist" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.133210 4770 scope.go:117] "RemoveContainer" containerID="5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131" Oct 04 04:27:11 crc kubenswrapper[4770]: E1004 04:27:11.133607 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131\": container with ID starting with 5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131 not found: ID does not exist" containerID="5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.133638 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131"} err="failed to get container status \"5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131\": rpc error: code = NotFound desc = could not find container \"5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131\": container with ID starting with 5504200ab4c7be1903e95d454b08402d3337fb27cee5284481729577d23f2131 not found: ID does not exist" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.171817 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6db4349c-ad5c-4141-a219-d1111575ab6a" (UID: "6db4349c-ad5c-4141-a219-d1111575ab6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.242470 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6db4349c-ad5c-4141-a219-d1111575ab6a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.336728 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-shhkz"] Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.336927 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-shhkz" podUID="c27542e5-b249-430d-8360-5453ccbe20c8" containerName="registry-server" containerID="cri-o://645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29" gracePeriod=2 Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.381546 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p2lnx"] Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.387581 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p2lnx"] Oct 04 04:27:11 crc kubenswrapper[4770]: I1004 04:27:11.694692 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6db4349c-ad5c-4141-a219-d1111575ab6a" path="/var/lib/kubelet/pods/6db4349c-ad5c-4141-a219-d1111575ab6a/volumes" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.040564 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.075797 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-catalog-content\") pod \"c27542e5-b249-430d-8360-5453ccbe20c8\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.075916 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bprbx\" (UniqueName: \"kubernetes.io/projected/c27542e5-b249-430d-8360-5453ccbe20c8-kube-api-access-bprbx\") pod \"c27542e5-b249-430d-8360-5453ccbe20c8\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.076109 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-utilities\") pod \"c27542e5-b249-430d-8360-5453ccbe20c8\" (UID: \"c27542e5-b249-430d-8360-5453ccbe20c8\") " Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.078155 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-utilities" (OuterVolumeSpecName: "utilities") pod "c27542e5-b249-430d-8360-5453ccbe20c8" (UID: "c27542e5-b249-430d-8360-5453ccbe20c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.084219 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c27542e5-b249-430d-8360-5453ccbe20c8-kube-api-access-bprbx" (OuterVolumeSpecName: "kube-api-access-bprbx") pod "c27542e5-b249-430d-8360-5453ccbe20c8" (UID: "c27542e5-b249-430d-8360-5453ccbe20c8"). InnerVolumeSpecName "kube-api-access-bprbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.084258 4770 generic.go:334] "Generic (PLEG): container finished" podID="c27542e5-b249-430d-8360-5453ccbe20c8" containerID="645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29" exitCode=0 Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.084289 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-shhkz" event={"ID":"c27542e5-b249-430d-8360-5453ccbe20c8","Type":"ContainerDied","Data":"645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29"} Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.084533 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-shhkz" event={"ID":"c27542e5-b249-430d-8360-5453ccbe20c8","Type":"ContainerDied","Data":"7e6cf514b2c55ad4c1721256e738361b4159e2642697d17e815c0678dfb45aaa"} Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.084563 4770 scope.go:117] "RemoveContainer" containerID="645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.084342 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-shhkz" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.139563 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c27542e5-b249-430d-8360-5453ccbe20c8" (UID: "c27542e5-b249-430d-8360-5453ccbe20c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.154795 4770 scope.go:117] "RemoveContainer" containerID="75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.174584 4770 scope.go:117] "RemoveContainer" containerID="c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.177899 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.177928 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27542e5-b249-430d-8360-5453ccbe20c8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.177942 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bprbx\" (UniqueName: \"kubernetes.io/projected/c27542e5-b249-430d-8360-5453ccbe20c8-kube-api-access-bprbx\") on node \"crc\" DevicePath \"\"" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.213909 4770 scope.go:117] "RemoveContainer" containerID="645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29" Oct 04 04:27:13 crc kubenswrapper[4770]: E1004 04:27:13.214374 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29\": container with ID starting with 645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29 not found: ID does not exist" containerID="645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.214412 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29"} err="failed to get container status \"645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29\": rpc error: code = NotFound desc = could not find container \"645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29\": container with ID starting with 645dc29e3d148840b4083155ced07b1577117dcabb657e49ca2d7b8bb69f7b29 not found: ID does not exist" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.214441 4770 scope.go:117] "RemoveContainer" containerID="75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a" Oct 04 04:27:13 crc kubenswrapper[4770]: E1004 04:27:13.214771 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a\": container with ID starting with 75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a not found: ID does not exist" containerID="75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.214800 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a"} err="failed to get container status \"75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a\": rpc error: code = NotFound desc = could not find container \"75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a\": container with ID starting with 75e7faa2cff417a1060fb33bcddab8d581d4e8fda333a14b3f73ee05b26bfd3a not found: ID does not exist" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.214821 4770 scope.go:117] "RemoveContainer" containerID="c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a" Oct 04 04:27:13 crc kubenswrapper[4770]: E1004 04:27:13.215269 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a\": container with ID starting with c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a not found: ID does not exist" containerID="c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.215314 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a"} err="failed to get container status \"c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a\": rpc error: code = NotFound desc = could not find container \"c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a\": container with ID starting with c33332a00fe48927cec92af8f0b8ba539faa09bdc365c0a7bd529e1161a2034a not found: ID does not exist" Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.412540 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-shhkz"] Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.417832 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-shhkz"] Oct 04 04:27:13 crc kubenswrapper[4770]: I1004 04:27:13.682343 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c27542e5-b249-430d-8360-5453ccbe20c8" path="/var/lib/kubelet/pods/c27542e5-b249-430d-8360-5453ccbe20c8/volumes" Oct 04 04:27:31 crc kubenswrapper[4770]: I1004 04:27:31.795650 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:27:31 crc kubenswrapper[4770]: I1004 04:27:31.796245 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:28:01 crc kubenswrapper[4770]: I1004 04:28:01.796747 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:28:01 crc kubenswrapper[4770]: I1004 04:28:01.797501 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:28:05 crc kubenswrapper[4770]: E1004 04:28:05.796780 4770 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="6.124s" Oct 04 04:28:05 crc kubenswrapper[4770]: I1004 04:28:05.797383 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:28:05 crc kubenswrapper[4770]: I1004 04:28:05.800599 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ee9dd0aa38f827331a89fde2cd36be108fb6bc2a67c9f1e927459a0a521b3c6"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:28:05 crc kubenswrapper[4770]: I1004 04:28:05.800755 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://6ee9dd0aa38f827331a89fde2cd36be108fb6bc2a67c9f1e927459a0a521b3c6" gracePeriod=600 Oct 04 04:28:08 crc kubenswrapper[4770]: I1004 04:28:08.682463 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="6ee9dd0aa38f827331a89fde2cd36be108fb6bc2a67c9f1e927459a0a521b3c6" exitCode=0 Oct 04 04:28:08 crc kubenswrapper[4770]: I1004 04:28:08.682568 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"6ee9dd0aa38f827331a89fde2cd36be108fb6bc2a67c9f1e927459a0a521b3c6"} Oct 04 04:28:08 crc kubenswrapper[4770]: I1004 04:28:08.683319 4770 scope.go:117] "RemoveContainer" containerID="7271edde77618cc741b6ff8681f027c192a58df17f7f3dd557efad82ee51bab4" Oct 04 04:28:09 crc kubenswrapper[4770]: I1004 04:28:09.707719 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5"} Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.054347 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 04 04:28:20 crc kubenswrapper[4770]: E1004 04:28:20.055640 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27542e5-b249-430d-8360-5453ccbe20c8" containerName="extract-utilities" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.055668 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27542e5-b249-430d-8360-5453ccbe20c8" containerName="extract-utilities" Oct 04 04:28:20 crc kubenswrapper[4770]: E1004 04:28:20.055701 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerName="registry-server" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.055715 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerName="registry-server" Oct 04 04:28:20 crc kubenswrapper[4770]: E1004 04:28:20.055740 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2436562a-3cf9-45df-862a-b9a969675e2d" containerName="extract-content" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.055755 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2436562a-3cf9-45df-862a-b9a969675e2d" containerName="extract-content" Oct 04 04:28:20 crc kubenswrapper[4770]: E1004 04:28:20.055781 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27542e5-b249-430d-8360-5453ccbe20c8" containerName="extract-content" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.055794 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27542e5-b249-430d-8360-5453ccbe20c8" containerName="extract-content" Oct 04 04:28:20 crc kubenswrapper[4770]: E1004 04:28:20.055817 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerName="extract-utilities" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.055830 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerName="extract-utilities" Oct 04 04:28:20 crc kubenswrapper[4770]: E1004 04:28:20.055851 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerName="extract-content" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.055864 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerName="extract-content" Oct 04 04:28:20 crc kubenswrapper[4770]: E1004 04:28:20.055889 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2436562a-3cf9-45df-862a-b9a969675e2d" containerName="extract-utilities" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.055902 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2436562a-3cf9-45df-862a-b9a969675e2d" containerName="extract-utilities" Oct 04 04:28:20 crc kubenswrapper[4770]: E1004 04:28:20.055925 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2436562a-3cf9-45df-862a-b9a969675e2d" containerName="registry-server" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.055938 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2436562a-3cf9-45df-862a-b9a969675e2d" containerName="registry-server" Oct 04 04:28:20 crc kubenswrapper[4770]: E1004 04:28:20.055959 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27542e5-b249-430d-8360-5453ccbe20c8" containerName="registry-server" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.055972 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27542e5-b249-430d-8360-5453ccbe20c8" containerName="registry-server" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.056342 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c27542e5-b249-430d-8360-5453ccbe20c8" containerName="registry-server" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.056375 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2436562a-3cf9-45df-862a-b9a969675e2d" containerName="registry-server" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.056418 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6db4349c-ad5c-4141-a219-d1111575ab6a" containerName="registry-server" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.057388 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.060510 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-t7x2l" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.064223 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.088936 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-83d8cec1-7494-446d-a411-cea543767bc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83d8cec1-7494-446d-a411-cea543767bc5\") pod \"mariadb-copy-data\" (UID: \"dcea39e7-a0fb-4831-81de-7cbecec5c963\") " pod="openstack/mariadb-copy-data" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.088986 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgzcg\" (UniqueName: \"kubernetes.io/projected/dcea39e7-a0fb-4831-81de-7cbecec5c963-kube-api-access-sgzcg\") pod \"mariadb-copy-data\" (UID: \"dcea39e7-a0fb-4831-81de-7cbecec5c963\") " pod="openstack/mariadb-copy-data" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.190400 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-83d8cec1-7494-446d-a411-cea543767bc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83d8cec1-7494-446d-a411-cea543767bc5\") pod \"mariadb-copy-data\" (UID: \"dcea39e7-a0fb-4831-81de-7cbecec5c963\") " pod="openstack/mariadb-copy-data" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.190914 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgzcg\" (UniqueName: \"kubernetes.io/projected/dcea39e7-a0fb-4831-81de-7cbecec5c963-kube-api-access-sgzcg\") pod \"mariadb-copy-data\" (UID: \"dcea39e7-a0fb-4831-81de-7cbecec5c963\") " pod="openstack/mariadb-copy-data" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.194977 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.195076 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-83d8cec1-7494-446d-a411-cea543767bc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83d8cec1-7494-446d-a411-cea543767bc5\") pod \"mariadb-copy-data\" (UID: \"dcea39e7-a0fb-4831-81de-7cbecec5c963\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/90896e6eade760bfbe3b5d6a6ec151e6fbd8af8d316e2fdfcec9cdeb118df65c/globalmount\"" pod="openstack/mariadb-copy-data" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.216768 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgzcg\" (UniqueName: \"kubernetes.io/projected/dcea39e7-a0fb-4831-81de-7cbecec5c963-kube-api-access-sgzcg\") pod \"mariadb-copy-data\" (UID: \"dcea39e7-a0fb-4831-81de-7cbecec5c963\") " pod="openstack/mariadb-copy-data" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.245902 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-83d8cec1-7494-446d-a411-cea543767bc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83d8cec1-7494-446d-a411-cea543767bc5\") pod \"mariadb-copy-data\" (UID: \"dcea39e7-a0fb-4831-81de-7cbecec5c963\") " pod="openstack/mariadb-copy-data" Oct 04 04:28:20 crc kubenswrapper[4770]: I1004 04:28:20.388113 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 04 04:28:21 crc kubenswrapper[4770]: I1004 04:28:21.013350 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 04 04:28:21 crc kubenswrapper[4770]: I1004 04:28:21.821457 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"dcea39e7-a0fb-4831-81de-7cbecec5c963","Type":"ContainerStarted","Data":"f4191c4e27e4d21d7679fd1172e3ac6c5627f44f8a756a032d06c4cf5c762683"} Oct 04 04:28:21 crc kubenswrapper[4770]: I1004 04:28:21.821676 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"dcea39e7-a0fb-4831-81de-7cbecec5c963","Type":"ContainerStarted","Data":"99a1ff1e2ad3c16d33bc882d2fc06fca73aac623f8d1867b26a5a7165c776ea7"} Oct 04 04:28:21 crc kubenswrapper[4770]: I1004 04:28:21.845850 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.84583788 podStartE2EDuration="2.84583788s" podCreationTimestamp="2025-10-04 04:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:28:21.842676467 +0000 UTC m=+5113.134686219" watchObservedRunningTime="2025-10-04 04:28:21.84583788 +0000 UTC m=+5113.137847592" Oct 04 04:28:23 crc kubenswrapper[4770]: I1004 04:28:23.797657 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:23 crc kubenswrapper[4770]: I1004 04:28:23.799663 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 04 04:28:23 crc kubenswrapper[4770]: I1004 04:28:23.809096 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:23 crc kubenswrapper[4770]: I1004 04:28:23.958069 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qs94\" (UniqueName: \"kubernetes.io/projected/a4b88cd7-6a91-41ea-b446-b4c4732d10ac-kube-api-access-6qs94\") pod \"mariadb-client\" (UID: \"a4b88cd7-6a91-41ea-b446-b4c4732d10ac\") " pod="openstack/mariadb-client" Oct 04 04:28:24 crc kubenswrapper[4770]: I1004 04:28:24.060696 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qs94\" (UniqueName: \"kubernetes.io/projected/a4b88cd7-6a91-41ea-b446-b4c4732d10ac-kube-api-access-6qs94\") pod \"mariadb-client\" (UID: \"a4b88cd7-6a91-41ea-b446-b4c4732d10ac\") " pod="openstack/mariadb-client" Oct 04 04:28:24 crc kubenswrapper[4770]: I1004 04:28:24.085107 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qs94\" (UniqueName: \"kubernetes.io/projected/a4b88cd7-6a91-41ea-b446-b4c4732d10ac-kube-api-access-6qs94\") pod \"mariadb-client\" (UID: \"a4b88cd7-6a91-41ea-b446-b4c4732d10ac\") " pod="openstack/mariadb-client" Oct 04 04:28:24 crc kubenswrapper[4770]: I1004 04:28:24.128566 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 04 04:28:24 crc kubenswrapper[4770]: I1004 04:28:24.336241 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:24 crc kubenswrapper[4770]: I1004 04:28:24.861246 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"a4b88cd7-6a91-41ea-b446-b4c4732d10ac","Type":"ContainerStarted","Data":"789012f634a9627cf83d8196e36f73bf24a89fc98031039d9beb22a95a285b05"} Oct 04 04:28:24 crc kubenswrapper[4770]: I1004 04:28:24.861320 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"a4b88cd7-6a91-41ea-b446-b4c4732d10ac","Type":"ContainerStarted","Data":"e14c186e025acd5f6ecbe7a405c24c20f57491a67be0edfeb1e5229d3def50fb"} Oct 04 04:28:25 crc kubenswrapper[4770]: I1004 04:28:25.872546 4770 generic.go:334] "Generic (PLEG): container finished" podID="a4b88cd7-6a91-41ea-b446-b4c4732d10ac" containerID="789012f634a9627cf83d8196e36f73bf24a89fc98031039d9beb22a95a285b05" exitCode=0 Oct 04 04:28:25 crc kubenswrapper[4770]: I1004 04:28:25.872632 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"a4b88cd7-6a91-41ea-b446-b4c4732d10ac","Type":"ContainerDied","Data":"789012f634a9627cf83d8196e36f73bf24a89fc98031039d9beb22a95a285b05"} Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.152316 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.179558 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_a4b88cd7-6a91-41ea-b446-b4c4732d10ac/mariadb-client/0.log" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.204982 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.210730 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.307434 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qs94\" (UniqueName: \"kubernetes.io/projected/a4b88cd7-6a91-41ea-b446-b4c4732d10ac-kube-api-access-6qs94\") pod \"a4b88cd7-6a91-41ea-b446-b4c4732d10ac\" (UID: \"a4b88cd7-6a91-41ea-b446-b4c4732d10ac\") " Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.314462 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4b88cd7-6a91-41ea-b446-b4c4732d10ac-kube-api-access-6qs94" (OuterVolumeSpecName: "kube-api-access-6qs94") pod "a4b88cd7-6a91-41ea-b446-b4c4732d10ac" (UID: "a4b88cd7-6a91-41ea-b446-b4c4732d10ac"). InnerVolumeSpecName "kube-api-access-6qs94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.343697 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:27 crc kubenswrapper[4770]: E1004 04:28:27.344245 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4b88cd7-6a91-41ea-b446-b4c4732d10ac" containerName="mariadb-client" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.344268 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4b88cd7-6a91-41ea-b446-b4c4732d10ac" containerName="mariadb-client" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.344691 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4b88cd7-6a91-41ea-b446-b4c4732d10ac" containerName="mariadb-client" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.345527 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.351240 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.409332 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qs94\" (UniqueName: \"kubernetes.io/projected/a4b88cd7-6a91-41ea-b446-b4c4732d10ac-kube-api-access-6qs94\") on node \"crc\" DevicePath \"\"" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.510861 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsc28\" (UniqueName: \"kubernetes.io/projected/73f03bb6-f3e1-4101-a819-8af0965209fb-kube-api-access-bsc28\") pod \"mariadb-client\" (UID: \"73f03bb6-f3e1-4101-a819-8af0965209fb\") " pod="openstack/mariadb-client" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.612524 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsc28\" (UniqueName: \"kubernetes.io/projected/73f03bb6-f3e1-4101-a819-8af0965209fb-kube-api-access-bsc28\") pod \"mariadb-client\" (UID: \"73f03bb6-f3e1-4101-a819-8af0965209fb\") " pod="openstack/mariadb-client" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.631330 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsc28\" (UniqueName: \"kubernetes.io/projected/73f03bb6-f3e1-4101-a819-8af0965209fb-kube-api-access-bsc28\") pod \"mariadb-client\" (UID: \"73f03bb6-f3e1-4101-a819-8af0965209fb\") " pod="openstack/mariadb-client" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.673561 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.698466 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4b88cd7-6a91-41ea-b446-b4c4732d10ac" path="/var/lib/kubelet/pods/a4b88cd7-6a91-41ea-b446-b4c4732d10ac/volumes" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.893126 4770 scope.go:117] "RemoveContainer" containerID="789012f634a9627cf83d8196e36f73bf24a89fc98031039d9beb22a95a285b05" Oct 04 04:28:27 crc kubenswrapper[4770]: I1004 04:28:27.893200 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 04 04:28:28 crc kubenswrapper[4770]: I1004 04:28:28.100593 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:28 crc kubenswrapper[4770]: W1004 04:28:28.103953 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73f03bb6_f3e1_4101_a819_8af0965209fb.slice/crio-ff10e27e2cb336422dfde221e0b6d2d8fd1a5f3bc735f0f40b18cb91645fac99 WatchSource:0}: Error finding container ff10e27e2cb336422dfde221e0b6d2d8fd1a5f3bc735f0f40b18cb91645fac99: Status 404 returned error can't find the container with id ff10e27e2cb336422dfde221e0b6d2d8fd1a5f3bc735f0f40b18cb91645fac99 Oct 04 04:28:28 crc kubenswrapper[4770]: I1004 04:28:28.902204 4770 generic.go:334] "Generic (PLEG): container finished" podID="73f03bb6-f3e1-4101-a819-8af0965209fb" containerID="cefd1e82582080ec5b190383f292f3f0ad42766d4594318ecf8e72a163b9a2ac" exitCode=0 Oct 04 04:28:28 crc kubenswrapper[4770]: I1004 04:28:28.902251 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"73f03bb6-f3e1-4101-a819-8af0965209fb","Type":"ContainerDied","Data":"cefd1e82582080ec5b190383f292f3f0ad42766d4594318ecf8e72a163b9a2ac"} Oct 04 04:28:28 crc kubenswrapper[4770]: I1004 04:28:28.902558 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"73f03bb6-f3e1-4101-a819-8af0965209fb","Type":"ContainerStarted","Data":"ff10e27e2cb336422dfde221e0b6d2d8fd1a5f3bc735f0f40b18cb91645fac99"} Oct 04 04:28:30 crc kubenswrapper[4770]: I1004 04:28:30.220211 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 04 04:28:30 crc kubenswrapper[4770]: I1004 04:28:30.242746 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_73f03bb6-f3e1-4101-a819-8af0965209fb/mariadb-client/0.log" Oct 04 04:28:30 crc kubenswrapper[4770]: I1004 04:28:30.309198 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:30 crc kubenswrapper[4770]: I1004 04:28:30.314312 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 04 04:28:30 crc kubenswrapper[4770]: I1004 04:28:30.357885 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsc28\" (UniqueName: \"kubernetes.io/projected/73f03bb6-f3e1-4101-a819-8af0965209fb-kube-api-access-bsc28\") pod \"73f03bb6-f3e1-4101-a819-8af0965209fb\" (UID: \"73f03bb6-f3e1-4101-a819-8af0965209fb\") " Oct 04 04:28:30 crc kubenswrapper[4770]: I1004 04:28:30.362845 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f03bb6-f3e1-4101-a819-8af0965209fb-kube-api-access-bsc28" (OuterVolumeSpecName: "kube-api-access-bsc28") pod "73f03bb6-f3e1-4101-a819-8af0965209fb" (UID: "73f03bb6-f3e1-4101-a819-8af0965209fb"). InnerVolumeSpecName "kube-api-access-bsc28". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:28:30 crc kubenswrapper[4770]: I1004 04:28:30.459510 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsc28\" (UniqueName: \"kubernetes.io/projected/73f03bb6-f3e1-4101-a819-8af0965209fb-kube-api-access-bsc28\") on node \"crc\" DevicePath \"\"" Oct 04 04:28:30 crc kubenswrapper[4770]: I1004 04:28:30.918953 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff10e27e2cb336422dfde221e0b6d2d8fd1a5f3bc735f0f40b18cb91645fac99" Oct 04 04:28:30 crc kubenswrapper[4770]: I1004 04:28:30.919037 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 04 04:28:31 crc kubenswrapper[4770]: I1004 04:28:31.682994 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f03bb6-f3e1-4101-a819-8af0965209fb" path="/var/lib/kubelet/pods/73f03bb6-f3e1-4101-a819-8af0965209fb/volumes" Oct 04 04:29:24 crc kubenswrapper[4770]: I1004 04:29:24.568911 4770 scope.go:117] "RemoveContainer" containerID="99d9238982da4cf11dc900c20fff1296210edfd18f43b37e905e5e5d354f4ebd" Oct 04 04:29:24 crc kubenswrapper[4770]: I1004 04:29:24.591813 4770 scope.go:117] "RemoveContainer" containerID="8191e38f4305ebb77c06af49baee1b27071748575b5c7db6966479d90b8e534e" Oct 04 04:29:24 crc kubenswrapper[4770]: I1004 04:29:24.626613 4770 scope.go:117] "RemoveContainer" containerID="232e4c7cfbfec43df63d8b53eba51f3c0fefad193e85529bc010f2ff6183b3c5" Oct 04 04:29:24 crc kubenswrapper[4770]: I1004 04:29:24.667653 4770 scope.go:117] "RemoveContainer" containerID="0380ddfb1198e43d4716d1111bd20230db09c0b5eded5ec6826d0222a67ddd86" Oct 04 04:29:31 crc kubenswrapper[4770]: E1004 04:29:31.631308 4770 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.68:35678->38.102.83.68:36801: write tcp 38.102.83.68:35678->38.102.83.68:36801: write: connection reset by peer Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.780885 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 04:29:35 crc kubenswrapper[4770]: E1004 04:29:35.781873 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f03bb6-f3e1-4101-a819-8af0965209fb" containerName="mariadb-client" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.781887 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f03bb6-f3e1-4101-a819-8af0965209fb" containerName="mariadb-client" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.782064 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="73f03bb6-f3e1-4101-a819-8af0965209fb" containerName="mariadb-client" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.782829 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.789341 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-kcll8" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.792379 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.792601 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.798228 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.799508 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.809326 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.811631 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.817463 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.823087 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.827495 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.946531 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfhkt\" (UniqueName: \"kubernetes.io/projected/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-kube-api-access-pfhkt\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.946813 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-820b5436-cbf5-40ad-8d30-d29ae9943ec5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-820b5436-cbf5-40ad-8d30-d29ae9943ec5\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.946922 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6254f2df-3ad1-4170-abbe-ba955e31bb82-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947027 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947122 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947195 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7e5b43-d3ed-42c0-8656-23460260ad8b-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947390 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-config\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947577 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df7e5b43-d3ed-42c0-8656-23460260ad8b-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947659 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df7e5b43-d3ed-42c0-8656-23460260ad8b-config\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947761 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-647zn\" (UniqueName: \"kubernetes.io/projected/6254f2df-3ad1-4170-abbe-ba955e31bb82-kube-api-access-647zn\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947809 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6254f2df-3ad1-4170-abbe-ba955e31bb82-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947850 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7aba1be6-57b4-4cfc-9f95-187e372c8b4c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7aba1be6-57b4-4cfc-9f95-187e372c8b4c\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.947934 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpsr9\" (UniqueName: \"kubernetes.io/projected/df7e5b43-d3ed-42c0-8656-23460260ad8b-kube-api-access-fpsr9\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.948045 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9b3a3dcb-c10f-4ca8-aa25-584c6ca38e73\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9b3a3dcb-c10f-4ca8-aa25-584c6ca38e73\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.948128 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6254f2df-3ad1-4170-abbe-ba955e31bb82-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.948329 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6254f2df-3ad1-4170-abbe-ba955e31bb82-config\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.948396 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df7e5b43-d3ed-42c0-8656-23460260ad8b-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.948438 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.984079 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.985810 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.988063 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-bmcv7" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.989156 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 04 04:29:35 crc kubenswrapper[4770]: I1004 04:29:35.989521 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.006625 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.008041 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.014632 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.027747 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.033851 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.035861 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.040793 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.049888 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpsr9\" (UniqueName: \"kubernetes.io/projected/df7e5b43-d3ed-42c0-8656-23460260ad8b-kube-api-access-fpsr9\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.049949 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9b3a3dcb-c10f-4ca8-aa25-584c6ca38e73\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9b3a3dcb-c10f-4ca8-aa25-584c6ca38e73\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.049981 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6254f2df-3ad1-4170-abbe-ba955e31bb82-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.050077 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6254f2df-3ad1-4170-abbe-ba955e31bb82-config\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.050105 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df7e5b43-d3ed-42c0-8656-23460260ad8b-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.050129 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.050158 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfhkt\" (UniqueName: \"kubernetes.io/projected/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-kube-api-access-pfhkt\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.050190 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-820b5436-cbf5-40ad-8d30-d29ae9943ec5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-820b5436-cbf5-40ad-8d30-d29ae9943ec5\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.050240 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6254f2df-3ad1-4170-abbe-ba955e31bb82-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.050291 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.050328 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.050352 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7e5b43-d3ed-42c0-8656-23460260ad8b-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.051109 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df7e5b43-d3ed-42c0-8656-23460260ad8b-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.051493 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.051888 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6254f2df-3ad1-4170-abbe-ba955e31bb82-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.052076 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-config\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.052141 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df7e5b43-d3ed-42c0-8656-23460260ad8b-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.052184 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df7e5b43-d3ed-42c0-8656-23460260ad8b-config\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.052235 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-647zn\" (UniqueName: \"kubernetes.io/projected/6254f2df-3ad1-4170-abbe-ba955e31bb82-kube-api-access-647zn\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.052276 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6254f2df-3ad1-4170-abbe-ba955e31bb82-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.052316 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7aba1be6-57b4-4cfc-9f95-187e372c8b4c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7aba1be6-57b4-4cfc-9f95-187e372c8b4c\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.052938 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.054215 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6254f2df-3ad1-4170-abbe-ba955e31bb82-config\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.069107 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpsr9\" (UniqueName: \"kubernetes.io/projected/df7e5b43-d3ed-42c0-8656-23460260ad8b-kube-api-access-fpsr9\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.074351 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df7e5b43-d3ed-42c0-8656-23460260ad8b-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.074603 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df7e5b43-d3ed-42c0-8656-23460260ad8b-config\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.074814 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6254f2df-3ad1-4170-abbe-ba955e31bb82-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.074970 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-config\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.075023 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.075173 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfhkt\" (UniqueName: \"kubernetes.io/projected/3aaa3b43-6349-4b31-a8fd-25bffad0fb69-kube-api-access-pfhkt\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.077803 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7e5b43-d3ed-42c0-8656-23460260ad8b-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.079440 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.079465 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.079486 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-820b5436-cbf5-40ad-8d30-d29ae9943ec5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-820b5436-cbf5-40ad-8d30-d29ae9943ec5\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7d1125e15bea50f58837212c74ecbf6533410cdc503618c5bae5e49f60a64154/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.079548 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.079488 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7aba1be6-57b4-4cfc-9f95-187e372c8b4c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7aba1be6-57b4-4cfc-9f95-187e372c8b4c\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/dc1211225a028324cf142aa50171bf187fcb9c0cd7646d537fcaee46df760e4f/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.079592 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9b3a3dcb-c10f-4ca8-aa25-584c6ca38e73\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9b3a3dcb-c10f-4ca8-aa25-584c6ca38e73\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ba9da1583a2dbcc876c4bed8eefea70a5c289324f343ce41591637a38cea4001/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.082088 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6254f2df-3ad1-4170-abbe-ba955e31bb82-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.092202 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-647zn\" (UniqueName: \"kubernetes.io/projected/6254f2df-3ad1-4170-abbe-ba955e31bb82-kube-api-access-647zn\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.108374 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-820b5436-cbf5-40ad-8d30-d29ae9943ec5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-820b5436-cbf5-40ad-8d30-d29ae9943ec5\") pod \"ovsdbserver-nb-1\" (UID: \"3aaa3b43-6349-4b31-a8fd-25bffad0fb69\") " pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.111252 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7aba1be6-57b4-4cfc-9f95-187e372c8b4c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7aba1be6-57b4-4cfc-9f95-187e372c8b4c\") pod \"ovsdbserver-nb-0\" (UID: \"6254f2df-3ad1-4170-abbe-ba955e31bb82\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.113106 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9b3a3dcb-c10f-4ca8-aa25-584c6ca38e73\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9b3a3dcb-c10f-4ca8-aa25-584c6ca38e73\") pod \"ovsdbserver-nb-2\" (UID: \"df7e5b43-d3ed-42c0-8656-23460260ad8b\") " pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.118947 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.146657 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.159454 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174374 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8jcx\" (UniqueName: \"kubernetes.io/projected/e7e0e8df-5568-462a-b40b-0ffc03b6197e-kube-api-access-h8jcx\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174404 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e7e0e8df-5568-462a-b40b-0ffc03b6197e-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174427 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e0e8df-5568-462a-b40b-0ffc03b6197e-config\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174450 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8c885a90-64e9-4c59-9e08-fbbfb07656ab-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174469 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/417c899d-895d-45ee-b4f2-5730df096041-config\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174496 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc7gf\" (UniqueName: \"kubernetes.io/projected/8c885a90-64e9-4c59-9e08-fbbfb07656ab-kube-api-access-nc7gf\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174513 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c885a90-64e9-4c59-9e08-fbbfb07656ab-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174533 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/417c899d-895d-45ee-b4f2-5730df096041-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174552 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7b01d4f6-51c1-4c4a-8c56-9d803046ae32\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b01d4f6-51c1-4c4a-8c56-9d803046ae32\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174572 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/417c899d-895d-45ee-b4f2-5730df096041-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174588 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7e0e8df-5568-462a-b40b-0ffc03b6197e-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174609 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/417c899d-895d-45ee-b4f2-5730df096041-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174630 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7q5m\" (UniqueName: \"kubernetes.io/projected/417c899d-895d-45ee-b4f2-5730df096041-kube-api-access-b7q5m\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174649 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0247d990-f1af-4ccf-97dc-08bfe5a237a7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0247d990-f1af-4ccf-97dc-08bfe5a237a7\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174677 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e0e8df-5568-462a-b40b-0ffc03b6197e-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174699 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c885a90-64e9-4c59-9e08-fbbfb07656ab-config\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174712 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c885a90-64e9-4c59-9e08-fbbfb07656ab-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.174735 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-09bf2244-7999-4bce-990a-9caf8931b2d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09bf2244-7999-4bce-990a-9caf8931b2d6\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.276189 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8jcx\" (UniqueName: \"kubernetes.io/projected/e7e0e8df-5568-462a-b40b-0ffc03b6197e-kube-api-access-h8jcx\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.276503 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e7e0e8df-5568-462a-b40b-0ffc03b6197e-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.276525 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e0e8df-5568-462a-b40b-0ffc03b6197e-config\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.276771 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8c885a90-64e9-4c59-9e08-fbbfb07656ab-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.276820 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/417c899d-895d-45ee-b4f2-5730df096041-config\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.276894 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc7gf\" (UniqueName: \"kubernetes.io/projected/8c885a90-64e9-4c59-9e08-fbbfb07656ab-kube-api-access-nc7gf\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.276919 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c885a90-64e9-4c59-9e08-fbbfb07656ab-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.276957 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/417c899d-895d-45ee-b4f2-5730df096041-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.276988 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7b01d4f6-51c1-4c4a-8c56-9d803046ae32\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b01d4f6-51c1-4c4a-8c56-9d803046ae32\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277031 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/417c899d-895d-45ee-b4f2-5730df096041-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277057 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7e0e8df-5568-462a-b40b-0ffc03b6197e-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277083 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/417c899d-895d-45ee-b4f2-5730df096041-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277125 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7q5m\" (UniqueName: \"kubernetes.io/projected/417c899d-895d-45ee-b4f2-5730df096041-kube-api-access-b7q5m\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277151 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0247d990-f1af-4ccf-97dc-08bfe5a237a7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0247d990-f1af-4ccf-97dc-08bfe5a237a7\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277219 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e0e8df-5568-462a-b40b-0ffc03b6197e-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277266 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c885a90-64e9-4c59-9e08-fbbfb07656ab-config\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277277 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e0e8df-5568-462a-b40b-0ffc03b6197e-config\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277280 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c885a90-64e9-4c59-9e08-fbbfb07656ab-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.277341 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-09bf2244-7999-4bce-990a-9caf8931b2d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09bf2244-7999-4bce-990a-9caf8931b2d6\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.278350 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e7e0e8df-5568-462a-b40b-0ffc03b6197e-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.279289 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/417c899d-895d-45ee-b4f2-5730df096041-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.279879 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c885a90-64e9-4c59-9e08-fbbfb07656ab-config\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.280206 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c885a90-64e9-4c59-9e08-fbbfb07656ab-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.280697 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.280748 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-09bf2244-7999-4bce-990a-9caf8931b2d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09bf2244-7999-4bce-990a-9caf8931b2d6\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1cf1694f1a76db010e6f725257b9cf049f0e6adee8c946d4bc1fb618b966d42f/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.281305 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/417c899d-895d-45ee-b4f2-5730df096041-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.281625 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.281662 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7b01d4f6-51c1-4c4a-8c56-9d803046ae32\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b01d4f6-51c1-4c4a-8c56-9d803046ae32\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/20ceea20edccce0a564550f8acc63be526b2e26064ef9c1d7f7470bb8353b06d/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.281686 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7e0e8df-5568-462a-b40b-0ffc03b6197e-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.281727 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.281771 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0247d990-f1af-4ccf-97dc-08bfe5a237a7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0247d990-f1af-4ccf-97dc-08bfe5a237a7\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bff3bdd8249bb0baeb909a916a30d4792a77a8eddbf66cad8f1036e788b4d659/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.282126 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8c885a90-64e9-4c59-9e08-fbbfb07656ab-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.282394 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/417c899d-895d-45ee-b4f2-5730df096041-config\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.284572 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/417c899d-895d-45ee-b4f2-5730df096041-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.286662 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e0e8df-5568-462a-b40b-0ffc03b6197e-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.290942 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c885a90-64e9-4c59-9e08-fbbfb07656ab-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.296948 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7q5m\" (UniqueName: \"kubernetes.io/projected/417c899d-895d-45ee-b4f2-5730df096041-kube-api-access-b7q5m\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.297406 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc7gf\" (UniqueName: \"kubernetes.io/projected/8c885a90-64e9-4c59-9e08-fbbfb07656ab-kube-api-access-nc7gf\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.298351 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8jcx\" (UniqueName: \"kubernetes.io/projected/e7e0e8df-5568-462a-b40b-0ffc03b6197e-kube-api-access-h8jcx\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.315524 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0247d990-f1af-4ccf-97dc-08bfe5a237a7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0247d990-f1af-4ccf-97dc-08bfe5a237a7\") pod \"ovsdbserver-sb-0\" (UID: \"417c899d-895d-45ee-b4f2-5730df096041\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.320222 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-09bf2244-7999-4bce-990a-9caf8931b2d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09bf2244-7999-4bce-990a-9caf8931b2d6\") pod \"ovsdbserver-sb-2\" (UID: \"e7e0e8df-5568-462a-b40b-0ffc03b6197e\") " pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.321437 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7b01d4f6-51c1-4c4a-8c56-9d803046ae32\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7b01d4f6-51c1-4c4a-8c56-9d803046ae32\") pod \"ovsdbserver-sb-1\" (UID: \"8c885a90-64e9-4c59-9e08-fbbfb07656ab\") " pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.326199 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.351349 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.607924 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.642055 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.749564 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 04 04:29:36 crc kubenswrapper[4770]: W1004 04:29:36.753898 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3aaa3b43_6349_4b31_a8fd_25bffad0fb69.slice/crio-1c20ac387aaca10122c5a57f07159add38d3b89f304bf95a82a3eb89b58c6e15 WatchSource:0}: Error finding container 1c20ac387aaca10122c5a57f07159add38d3b89f304bf95a82a3eb89b58c6e15: Status 404 returned error can't find the container with id 1c20ac387aaca10122c5a57f07159add38d3b89f304bf95a82a3eb89b58c6e15 Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.841400 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 04 04:29:36 crc kubenswrapper[4770]: W1004 04:29:36.855216 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf7e5b43_d3ed_42c0_8656_23460260ad8b.slice/crio-175861b65aa428d8323d717c1b63ab404f60d35bd355c44fd1b230af93def0b3 WatchSource:0}: Error finding container 175861b65aa428d8323d717c1b63ab404f60d35bd355c44fd1b230af93def0b3: Status 404 returned error can't find the container with id 175861b65aa428d8323d717c1b63ab404f60d35bd355c44fd1b230af93def0b3 Oct 04 04:29:36 crc kubenswrapper[4770]: W1004 04:29:36.940383 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7e0e8df_5568_462a_b40b_0ffc03b6197e.slice/crio-743af6d4121218013ea248a6570b09bfe13c5cc4b6de2d0fa3d12af78cddc117 WatchSource:0}: Error finding container 743af6d4121218013ea248a6570b09bfe13c5cc4b6de2d0fa3d12af78cddc117: Status 404 returned error can't find the container with id 743af6d4121218013ea248a6570b09bfe13c5cc4b6de2d0fa3d12af78cddc117 Oct 04 04:29:36 crc kubenswrapper[4770]: I1004 04:29:36.945114 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.192037 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 04:29:37 crc kubenswrapper[4770]: W1004 04:29:37.204367 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod417c899d_895d_45ee_b4f2_5730df096041.slice/crio-faf384e44790ca55e8ed045b3002d44c5ee33b12a7fb4cdf240a2d97a7cb8094 WatchSource:0}: Error finding container faf384e44790ca55e8ed045b3002d44c5ee33b12a7fb4cdf240a2d97a7cb8094: Status 404 returned error can't find the container with id faf384e44790ca55e8ed045b3002d44c5ee33b12a7fb4cdf240a2d97a7cb8094 Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.427058 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 04 04:29:37 crc kubenswrapper[4770]: W1004 04:29:37.437177 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c885a90_64e9_4c59_9e08_fbbfb07656ab.slice/crio-edc55dd5e0ba4cdcdb498f4aa1f58d3b4ac8373ee58d8681f67712939df3df11 WatchSource:0}: Error finding container edc55dd5e0ba4cdcdb498f4aa1f58d3b4ac8373ee58d8681f67712939df3df11: Status 404 returned error can't find the container with id edc55dd5e0ba4cdcdb498f4aa1f58d3b4ac8373ee58d8681f67712939df3df11 Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.468634 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"8c885a90-64e9-4c59-9e08-fbbfb07656ab","Type":"ContainerStarted","Data":"edc55dd5e0ba4cdcdb498f4aa1f58d3b4ac8373ee58d8681f67712939df3df11"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.471739 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"417c899d-895d-45ee-b4f2-5730df096041","Type":"ContainerStarted","Data":"debf36d117ca59988503f3f0cb851e6bbf6d8ab19ef7f3b18c83ca3fb2fc35c7"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.471792 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"417c899d-895d-45ee-b4f2-5730df096041","Type":"ContainerStarted","Data":"faf384e44790ca55e8ed045b3002d44c5ee33b12a7fb4cdf240a2d97a7cb8094"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.474537 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"e7e0e8df-5568-462a-b40b-0ffc03b6197e","Type":"ContainerStarted","Data":"50fef77e63619acc73f8b248621fa95a664edeb5b643d0dc37bd0171a790411b"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.474570 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"e7e0e8df-5568-462a-b40b-0ffc03b6197e","Type":"ContainerStarted","Data":"0e3d896f077561a563a0f36ee6cbecb1c3211d1769eff1c4c10653a5446f9479"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.474583 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"e7e0e8df-5568-462a-b40b-0ffc03b6197e","Type":"ContainerStarted","Data":"743af6d4121218013ea248a6570b09bfe13c5cc4b6de2d0fa3d12af78cddc117"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.477349 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"df7e5b43-d3ed-42c0-8656-23460260ad8b","Type":"ContainerStarted","Data":"c3097607722680062f70873db39abca0be27b2e28124f7e5ea089d0f5729c264"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.477381 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"df7e5b43-d3ed-42c0-8656-23460260ad8b","Type":"ContainerStarted","Data":"d0ce669719d2384d34b95adf8f47e0866db8bada6505168d73f05d9cef32d6e9"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.477395 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"df7e5b43-d3ed-42c0-8656-23460260ad8b","Type":"ContainerStarted","Data":"175861b65aa428d8323d717c1b63ab404f60d35bd355c44fd1b230af93def0b3"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.480656 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6254f2df-3ad1-4170-abbe-ba955e31bb82","Type":"ContainerStarted","Data":"088b53a168cfe99f98cec8657193bf21f90bc03c76024fbdf36d457edd32439e"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.480694 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6254f2df-3ad1-4170-abbe-ba955e31bb82","Type":"ContainerStarted","Data":"1112c3585273aa3d101a5d705f745f44f98303e8e76436c18f41e4aa380be9e5"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.480708 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6254f2df-3ad1-4170-abbe-ba955e31bb82","Type":"ContainerStarted","Data":"87657f5fadd15fc01dab53b122cbd3d10b926c3474086409c5a83d5d127ada8c"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.482456 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"3aaa3b43-6349-4b31-a8fd-25bffad0fb69","Type":"ContainerStarted","Data":"ed40ff76bc9b3a9add9f946123c97916b2a138bb9ab5bd0dd1fa55fbdc334303"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.482483 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"3aaa3b43-6349-4b31-a8fd-25bffad0fb69","Type":"ContainerStarted","Data":"4908b306ff29f078be4081ddbf0040ad54e6c87d0010695a5930b9e75f108d9d"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.482496 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"3aaa3b43-6349-4b31-a8fd-25bffad0fb69","Type":"ContainerStarted","Data":"1c20ac387aaca10122c5a57f07159add38d3b89f304bf95a82a3eb89b58c6e15"} Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.496168 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.496148954 podStartE2EDuration="3.496148954s" podCreationTimestamp="2025-10-04 04:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:29:37.491231415 +0000 UTC m=+5188.783241127" watchObservedRunningTime="2025-10-04 04:29:37.496148954 +0000 UTC m=+5188.788158666" Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.515790 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.515771267 podStartE2EDuration="3.515771267s" podCreationTimestamp="2025-10-04 04:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:29:37.509425861 +0000 UTC m=+5188.801435573" watchObservedRunningTime="2025-10-04 04:29:37.515771267 +0000 UTC m=+5188.807780979" Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.532565 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.532546216 podStartE2EDuration="3.532546216s" podCreationTimestamp="2025-10-04 04:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:29:37.527272278 +0000 UTC m=+5188.819281990" watchObservedRunningTime="2025-10-04 04:29:37.532546216 +0000 UTC m=+5188.824555928" Oct 04 04:29:37 crc kubenswrapper[4770]: I1004 04:29:37.549869 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.549848968 podStartE2EDuration="3.549848968s" podCreationTimestamp="2025-10-04 04:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:29:37.542958267 +0000 UTC m=+5188.834967979" watchObservedRunningTime="2025-10-04 04:29:37.549848968 +0000 UTC m=+5188.841858700" Oct 04 04:29:38 crc kubenswrapper[4770]: I1004 04:29:38.498900 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"417c899d-895d-45ee-b4f2-5730df096041","Type":"ContainerStarted","Data":"bad7d681cb00a0ca043f19fbde7b6a1b9ccbf956a2be655b001c0e0440d76eba"} Oct 04 04:29:38 crc kubenswrapper[4770]: I1004 04:29:38.501934 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"8c885a90-64e9-4c59-9e08-fbbfb07656ab","Type":"ContainerStarted","Data":"d5541c1cf78519ef56692aa9e40cf1e4f8a22d8382e21da0f77f31fbbd1e4ff9"} Oct 04 04:29:38 crc kubenswrapper[4770]: I1004 04:29:38.501973 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"8c885a90-64e9-4c59-9e08-fbbfb07656ab","Type":"ContainerStarted","Data":"2d52c32dbae5cc0e17b24c88ea4a9a627abfa02b619ef97a2ffa351edfeba36b"} Oct 04 04:29:38 crc kubenswrapper[4770]: I1004 04:29:38.516609 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.516586965 podStartE2EDuration="4.516586965s" podCreationTimestamp="2025-10-04 04:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:29:38.513555216 +0000 UTC m=+5189.805564948" watchObservedRunningTime="2025-10-04 04:29:38.516586965 +0000 UTC m=+5189.808596687" Oct 04 04:29:38 crc kubenswrapper[4770]: I1004 04:29:38.539878 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.5398510420000004 podStartE2EDuration="4.539851042s" podCreationTimestamp="2025-10-04 04:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:29:38.534184934 +0000 UTC m=+5189.826194666" watchObservedRunningTime="2025-10-04 04:29:38.539851042 +0000 UTC m=+5189.831860754" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.119395 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.147458 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.160134 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.173244 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.189318 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.202579 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.328498 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.352426 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.508970 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.509047 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.509064 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:39 crc kubenswrapper[4770]: I1004 04:29:39.609082 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.182119 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.228404 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.228814 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.327207 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.353107 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.419094 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6fd8785b9-r2w78"] Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.420658 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.422661 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.438165 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fd8785b9-r2w78"] Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.575288 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-config\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.575335 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-ovsdbserver-nb\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.575537 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5kvw\" (UniqueName: \"kubernetes.io/projected/f345f864-05d1-46b3-b41a-2af73b2d8c64-kube-api-access-w5kvw\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.575835 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-dns-svc\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.609155 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.677243 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5kvw\" (UniqueName: \"kubernetes.io/projected/f345f864-05d1-46b3-b41a-2af73b2d8c64-kube-api-access-w5kvw\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.677355 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-dns-svc\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.677438 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-config\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.677459 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-ovsdbserver-nb\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.678435 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-dns-svc\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.679247 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-ovsdbserver-nb\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.679666 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-config\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.698614 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5kvw\" (UniqueName: \"kubernetes.io/projected/f345f864-05d1-46b3-b41a-2af73b2d8c64-kube-api-access-w5kvw\") pod \"dnsmasq-dns-6fd8785b9-r2w78\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:41 crc kubenswrapper[4770]: I1004 04:29:41.740980 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.181463 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fd8785b9-r2w78"] Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.365604 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.388313 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.411896 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.541804 4770 generic.go:334] "Generic (PLEG): container finished" podID="f345f864-05d1-46b3-b41a-2af73b2d8c64" containerID="f7ade7e5774951aa70178dbd6a25a6da0f5dbef79302330caae5301b1234ae2f" exitCode=0 Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.541864 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" event={"ID":"f345f864-05d1-46b3-b41a-2af73b2d8c64","Type":"ContainerDied","Data":"f7ade7e5774951aa70178dbd6a25a6da0f5dbef79302330caae5301b1234ae2f"} Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.541923 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" event={"ID":"f345f864-05d1-46b3-b41a-2af73b2d8c64","Type":"ContainerStarted","Data":"2e7f3f78f1f5ff3c651b294a276387c2ae5b752fc2de2341ea5f5dbdc7cf775d"} Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.596471 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.684565 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fd8785b9-r2w78"] Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.688831 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.763487 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-778dbf57b9-d5qp2"] Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.765434 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.766722 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-778dbf57b9-d5qp2"] Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.768572 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.768791 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.807865 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-sb\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.807905 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-config\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.807945 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-dns-svc\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.808023 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-nb\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.808094 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twz8r\" (UniqueName: \"kubernetes.io/projected/381071ee-0780-4ba9-a924-4da9ea4113b2-kube-api-access-twz8r\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.909708 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-nb\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.910032 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twz8r\" (UniqueName: \"kubernetes.io/projected/381071ee-0780-4ba9-a924-4da9ea4113b2-kube-api-access-twz8r\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.910064 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-sb\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.910081 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-config\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.910116 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-dns-svc\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.911024 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-dns-svc\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.911047 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-config\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.911539 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-sb\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.911556 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-nb\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:42 crc kubenswrapper[4770]: I1004 04:29:42.939948 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twz8r\" (UniqueName: \"kubernetes.io/projected/381071ee-0780-4ba9-a924-4da9ea4113b2-kube-api-access-twz8r\") pod \"dnsmasq-dns-778dbf57b9-d5qp2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:43 crc kubenswrapper[4770]: I1004 04:29:43.081267 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:43 crc kubenswrapper[4770]: I1004 04:29:43.513878 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-778dbf57b9-d5qp2"] Oct 04 04:29:43 crc kubenswrapper[4770]: I1004 04:29:43.551939 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" event={"ID":"f345f864-05d1-46b3-b41a-2af73b2d8c64","Type":"ContainerStarted","Data":"4efa024db728ef84763468355c2e43f41791002f10d9b2e2b1363c5d5a033ea0"} Oct 04 04:29:43 crc kubenswrapper[4770]: I1004 04:29:43.552476 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" podUID="f345f864-05d1-46b3-b41a-2af73b2d8c64" containerName="dnsmasq-dns" containerID="cri-o://4efa024db728ef84763468355c2e43f41791002f10d9b2e2b1363c5d5a033ea0" gracePeriod=10 Oct 04 04:29:43 crc kubenswrapper[4770]: I1004 04:29:43.552978 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:43 crc kubenswrapper[4770]: I1004 04:29:43.554934 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" event={"ID":"381071ee-0780-4ba9-a924-4da9ea4113b2","Type":"ContainerStarted","Data":"57a1f4d7720ef139005e93af488ea6be11f74b5666da63fd8889069313ebc829"} Oct 04 04:29:43 crc kubenswrapper[4770]: I1004 04:29:43.582349 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" podStartSLOduration=2.582307373 podStartE2EDuration="2.582307373s" podCreationTimestamp="2025-10-04 04:29:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:29:43.578301928 +0000 UTC m=+5194.870311640" watchObservedRunningTime="2025-10-04 04:29:43.582307373 +0000 UTC m=+5194.874317085" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.565214 4770 generic.go:334] "Generic (PLEG): container finished" podID="f345f864-05d1-46b3-b41a-2af73b2d8c64" containerID="4efa024db728ef84763468355c2e43f41791002f10d9b2e2b1363c5d5a033ea0" exitCode=0 Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.565244 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" event={"ID":"f345f864-05d1-46b3-b41a-2af73b2d8c64","Type":"ContainerDied","Data":"4efa024db728ef84763468355c2e43f41791002f10d9b2e2b1363c5d5a033ea0"} Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.565656 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" event={"ID":"f345f864-05d1-46b3-b41a-2af73b2d8c64","Type":"ContainerDied","Data":"2e7f3f78f1f5ff3c651b294a276387c2ae5b752fc2de2341ea5f5dbdc7cf775d"} Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.565674 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e7f3f78f1f5ff3c651b294a276387c2ae5b752fc2de2341ea5f5dbdc7cf775d" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.568756 4770 generic.go:334] "Generic (PLEG): container finished" podID="381071ee-0780-4ba9-a924-4da9ea4113b2" containerID="123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86" exitCode=0 Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.568801 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" event={"ID":"381071ee-0780-4ba9-a924-4da9ea4113b2","Type":"ContainerDied","Data":"123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86"} Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.649222 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.842161 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-ovsdbserver-nb\") pod \"f345f864-05d1-46b3-b41a-2af73b2d8c64\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.842279 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5kvw\" (UniqueName: \"kubernetes.io/projected/f345f864-05d1-46b3-b41a-2af73b2d8c64-kube-api-access-w5kvw\") pod \"f345f864-05d1-46b3-b41a-2af73b2d8c64\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.842326 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-config\") pod \"f345f864-05d1-46b3-b41a-2af73b2d8c64\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.842392 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-dns-svc\") pod \"f345f864-05d1-46b3-b41a-2af73b2d8c64\" (UID: \"f345f864-05d1-46b3-b41a-2af73b2d8c64\") " Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.859186 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f345f864-05d1-46b3-b41a-2af73b2d8c64-kube-api-access-w5kvw" (OuterVolumeSpecName: "kube-api-access-w5kvw") pod "f345f864-05d1-46b3-b41a-2af73b2d8c64" (UID: "f345f864-05d1-46b3-b41a-2af73b2d8c64"). InnerVolumeSpecName "kube-api-access-w5kvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.878785 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f345f864-05d1-46b3-b41a-2af73b2d8c64" (UID: "f345f864-05d1-46b3-b41a-2af73b2d8c64"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.879225 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-config" (OuterVolumeSpecName: "config") pod "f345f864-05d1-46b3-b41a-2af73b2d8c64" (UID: "f345f864-05d1-46b3-b41a-2af73b2d8c64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.887653 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f345f864-05d1-46b3-b41a-2af73b2d8c64" (UID: "f345f864-05d1-46b3-b41a-2af73b2d8c64"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.944466 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.944498 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.944510 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5kvw\" (UniqueName: \"kubernetes.io/projected/f345f864-05d1-46b3-b41a-2af73b2d8c64-kube-api-access-w5kvw\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:44 crc kubenswrapper[4770]: I1004 04:29:44.944519 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f345f864-05d1-46b3-b41a-2af73b2d8c64-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.553349 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 04 04:29:45 crc kubenswrapper[4770]: E1004 04:29:45.554303 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f345f864-05d1-46b3-b41a-2af73b2d8c64" containerName="init" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.554327 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f345f864-05d1-46b3-b41a-2af73b2d8c64" containerName="init" Oct 04 04:29:45 crc kubenswrapper[4770]: E1004 04:29:45.554340 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f345f864-05d1-46b3-b41a-2af73b2d8c64" containerName="dnsmasq-dns" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.554349 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f345f864-05d1-46b3-b41a-2af73b2d8c64" containerName="dnsmasq-dns" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.554693 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f345f864-05d1-46b3-b41a-2af73b2d8c64" containerName="dnsmasq-dns" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.555435 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.560937 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.579202 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.584429 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fd8785b9-r2w78" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.584426 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" event={"ID":"381071ee-0780-4ba9-a924-4da9ea4113b2","Type":"ContainerStarted","Data":"958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a"} Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.627322 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" podStartSLOduration=3.62729725 podStartE2EDuration="3.62729725s" podCreationTimestamp="2025-10-04 04:29:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:29:45.615667226 +0000 UTC m=+5196.907676938" watchObservedRunningTime="2025-10-04 04:29:45.62729725 +0000 UTC m=+5196.919306992" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.640256 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fd8785b9-r2w78"] Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.647075 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6fd8785b9-r2w78"] Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.685827 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f345f864-05d1-46b3-b41a-2af73b2d8c64" path="/var/lib/kubelet/pods/f345f864-05d1-46b3-b41a-2af73b2d8c64/volumes" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.757529 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.757765 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/b7daf327-2122-4ef8-8352-c41822f77200-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.757803 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk4nc\" (UniqueName: \"kubernetes.io/projected/b7daf327-2122-4ef8-8352-c41822f77200-kube-api-access-pk4nc\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.859590 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/b7daf327-2122-4ef8-8352-c41822f77200-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.859658 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk4nc\" (UniqueName: \"kubernetes.io/projected/b7daf327-2122-4ef8-8352-c41822f77200-kube-api-access-pk4nc\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.859717 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.864316 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.864510 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9c1e9a8fd20a4f715a8a60362f6894057b51287ec96e72c3cd9681d8321321a9/globalmount\"" pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.874567 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/b7daf327-2122-4ef8-8352-c41822f77200-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.880834 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk4nc\" (UniqueName: \"kubernetes.io/projected/b7daf327-2122-4ef8-8352-c41822f77200-kube-api-access-pk4nc\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " pod="openstack/ovn-copy-data" Oct 04 04:29:45 crc kubenswrapper[4770]: I1004 04:29:45.905130 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\") pod \"ovn-copy-data\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " pod="openstack/ovn-copy-data" Oct 04 04:29:46 crc kubenswrapper[4770]: I1004 04:29:46.174066 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 04 04:29:46 crc kubenswrapper[4770]: I1004 04:29:46.596168 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:46 crc kubenswrapper[4770]: I1004 04:29:46.680581 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 04 04:29:46 crc kubenswrapper[4770]: W1004 04:29:46.683545 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7daf327_2122_4ef8_8352_c41822f77200.slice/crio-3aff780f573332b464e041c318c5d5a034e4308f77f985ca41979a55542f25f3 WatchSource:0}: Error finding container 3aff780f573332b464e041c318c5d5a034e4308f77f985ca41979a55542f25f3: Status 404 returned error can't find the container with id 3aff780f573332b464e041c318c5d5a034e4308f77f985ca41979a55542f25f3 Oct 04 04:29:46 crc kubenswrapper[4770]: I1004 04:29:46.687304 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:29:47 crc kubenswrapper[4770]: I1004 04:29:47.607131 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"b7daf327-2122-4ef8-8352-c41822f77200","Type":"ContainerStarted","Data":"3aff780f573332b464e041c318c5d5a034e4308f77f985ca41979a55542f25f3"} Oct 04 04:29:48 crc kubenswrapper[4770]: I1004 04:29:48.617896 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"b7daf327-2122-4ef8-8352-c41822f77200","Type":"ContainerStarted","Data":"3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec"} Oct 04 04:29:48 crc kubenswrapper[4770]: I1004 04:29:48.637197 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.785681223 podStartE2EDuration="4.637179207s" podCreationTimestamp="2025-10-04 04:29:44 +0000 UTC" firstStartedPulling="2025-10-04 04:29:46.68712071 +0000 UTC m=+5197.979130412" lastFinishedPulling="2025-10-04 04:29:47.538618674 +0000 UTC m=+5198.830628396" observedRunningTime="2025-10-04 04:29:48.630559033 +0000 UTC m=+5199.922568745" watchObservedRunningTime="2025-10-04 04:29:48.637179207 +0000 UTC m=+5199.929188909" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.083136 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.149875 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b6b49df6f-vpzhl"] Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.150270 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" podUID="89738376-3973-41e0-9ee4-b9157c3c535a" containerName="dnsmasq-dns" containerID="cri-o://c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b" gracePeriod=10 Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.621997 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.659735 4770 generic.go:334] "Generic (PLEG): container finished" podID="89738376-3973-41e0-9ee4-b9157c3c535a" containerID="c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b" exitCode=0 Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.659772 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" event={"ID":"89738376-3973-41e0-9ee4-b9157c3c535a","Type":"ContainerDied","Data":"c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b"} Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.659794 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" event={"ID":"89738376-3973-41e0-9ee4-b9157c3c535a","Type":"ContainerDied","Data":"eddbd90fdad53719497c5f3e710e83fb82a7e8edbb7b1d776050eed6e33d34ff"} Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.659810 4770 scope.go:117] "RemoveContainer" containerID="c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.659914 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6b49df6f-vpzhl" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.698194 4770 scope.go:117] "RemoveContainer" containerID="10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.726735 4770 scope.go:117] "RemoveContainer" containerID="c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b" Oct 04 04:29:53 crc kubenswrapper[4770]: E1004 04:29:53.727231 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b\": container with ID starting with c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b not found: ID does not exist" containerID="c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.727269 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b"} err="failed to get container status \"c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b\": rpc error: code = NotFound desc = could not find container \"c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b\": container with ID starting with c981b93a9631a2225452d52b6773d32470c6cf3f5498cf0cc9dbe30fc93f6c4b not found: ID does not exist" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.727294 4770 scope.go:117] "RemoveContainer" containerID="10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a" Oct 04 04:29:53 crc kubenswrapper[4770]: E1004 04:29:53.727683 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a\": container with ID starting with 10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a not found: ID does not exist" containerID="10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.727712 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a"} err="failed to get container status \"10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a\": rpc error: code = NotFound desc = could not find container \"10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a\": container with ID starting with 10778f37cf449acc8cd1f164b831336900cf26b7c5138b4a216350cef6ed969a not found: ID does not exist" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.800886 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7zkk\" (UniqueName: \"kubernetes.io/projected/89738376-3973-41e0-9ee4-b9157c3c535a-kube-api-access-j7zkk\") pod \"89738376-3973-41e0-9ee4-b9157c3c535a\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.801048 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-dns-svc\") pod \"89738376-3973-41e0-9ee4-b9157c3c535a\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.801154 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-config\") pod \"89738376-3973-41e0-9ee4-b9157c3c535a\" (UID: \"89738376-3973-41e0-9ee4-b9157c3c535a\") " Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.815257 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89738376-3973-41e0-9ee4-b9157c3c535a-kube-api-access-j7zkk" (OuterVolumeSpecName: "kube-api-access-j7zkk") pod "89738376-3973-41e0-9ee4-b9157c3c535a" (UID: "89738376-3973-41e0-9ee4-b9157c3c535a"). InnerVolumeSpecName "kube-api-access-j7zkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.837777 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "89738376-3973-41e0-9ee4-b9157c3c535a" (UID: "89738376-3973-41e0-9ee4-b9157c3c535a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.839777 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-config" (OuterVolumeSpecName: "config") pod "89738376-3973-41e0-9ee4-b9157c3c535a" (UID: "89738376-3973-41e0-9ee4-b9157c3c535a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.902573 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7zkk\" (UniqueName: \"kubernetes.io/projected/89738376-3973-41e0-9ee4-b9157c3c535a-kube-api-access-j7zkk\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.902610 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:53 crc kubenswrapper[4770]: I1004 04:29:53.902621 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89738376-3973-41e0-9ee4-b9157c3c535a-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.002190 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b6b49df6f-vpzhl"] Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.009629 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b6b49df6f-vpzhl"] Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.303403 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 04 04:29:54 crc kubenswrapper[4770]: E1004 04:29:54.303937 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89738376-3973-41e0-9ee4-b9157c3c535a" containerName="init" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.303950 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="89738376-3973-41e0-9ee4-b9157c3c535a" containerName="init" Oct 04 04:29:54 crc kubenswrapper[4770]: E1004 04:29:54.303963 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89738376-3973-41e0-9ee4-b9157c3c535a" containerName="dnsmasq-dns" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.303969 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="89738376-3973-41e0-9ee4-b9157c3c535a" containerName="dnsmasq-dns" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.304160 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="89738376-3973-41e0-9ee4-b9157c3c535a" containerName="dnsmasq-dns" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.305065 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.311309 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-mkgm4" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.311611 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.334791 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.340456 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.413911 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fbbe2b0e-dad0-4407-9347-59447104e584-scripts\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.413982 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv9ks\" (UniqueName: \"kubernetes.io/projected/fbbe2b0e-dad0-4407-9347-59447104e584-kube-api-access-zv9ks\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.414051 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbbe2b0e-dad0-4407-9347-59447104e584-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.414075 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbe2b0e-dad0-4407-9347-59447104e584-config\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.414105 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fbbe2b0e-dad0-4407-9347-59447104e584-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.515713 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fbbe2b0e-dad0-4407-9347-59447104e584-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.515831 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fbbe2b0e-dad0-4407-9347-59447104e584-scripts\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.515855 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv9ks\" (UniqueName: \"kubernetes.io/projected/fbbe2b0e-dad0-4407-9347-59447104e584-kube-api-access-zv9ks\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.515885 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbbe2b0e-dad0-4407-9347-59447104e584-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.515901 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbe2b0e-dad0-4407-9347-59447104e584-config\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.516374 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fbbe2b0e-dad0-4407-9347-59447104e584-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.516693 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbe2b0e-dad0-4407-9347-59447104e584-config\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.517994 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fbbe2b0e-dad0-4407-9347-59447104e584-scripts\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.526035 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbbe2b0e-dad0-4407-9347-59447104e584-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.550227 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv9ks\" (UniqueName: \"kubernetes.io/projected/fbbe2b0e-dad0-4407-9347-59447104e584-kube-api-access-zv9ks\") pod \"ovn-northd-0\" (UID: \"fbbe2b0e-dad0-4407-9347-59447104e584\") " pod="openstack/ovn-northd-0" Oct 04 04:29:54 crc kubenswrapper[4770]: I1004 04:29:54.668850 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 04:29:55 crc kubenswrapper[4770]: W1004 04:29:55.103244 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbbe2b0e_dad0_4407_9347_59447104e584.slice/crio-c7559507b1e94baa28e77324b8248b4e3631a23c449e50d8f3d07eca455894c5 WatchSource:0}: Error finding container c7559507b1e94baa28e77324b8248b4e3631a23c449e50d8f3d07eca455894c5: Status 404 returned error can't find the container with id c7559507b1e94baa28e77324b8248b4e3631a23c449e50d8f3d07eca455894c5 Oct 04 04:29:55 crc kubenswrapper[4770]: I1004 04:29:55.103867 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 04:29:55 crc kubenswrapper[4770]: I1004 04:29:55.686870 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89738376-3973-41e0-9ee4-b9157c3c535a" path="/var/lib/kubelet/pods/89738376-3973-41e0-9ee4-b9157c3c535a/volumes" Oct 04 04:29:55 crc kubenswrapper[4770]: I1004 04:29:55.688297 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fbbe2b0e-dad0-4407-9347-59447104e584","Type":"ContainerStarted","Data":"f315b89434b94d1e0459464a0fdff909724487ecf64415797f008411d5cb64f8"} Oct 04 04:29:55 crc kubenswrapper[4770]: I1004 04:29:55.688340 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fbbe2b0e-dad0-4407-9347-59447104e584","Type":"ContainerStarted","Data":"c7559507b1e94baa28e77324b8248b4e3631a23c449e50d8f3d07eca455894c5"} Oct 04 04:29:56 crc kubenswrapper[4770]: I1004 04:29:56.690156 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fbbe2b0e-dad0-4407-9347-59447104e584","Type":"ContainerStarted","Data":"54ad52e98f5c2638d44752dee7a29e2f33c040233dc557e8043684dfb7165ffc"} Oct 04 04:29:56 crc kubenswrapper[4770]: I1004 04:29:56.690768 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 04 04:29:56 crc kubenswrapper[4770]: I1004 04:29:56.713578 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.713558182 podStartE2EDuration="2.713558182s" podCreationTimestamp="2025-10-04 04:29:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:29:56.709645509 +0000 UTC m=+5208.001655251" watchObservedRunningTime="2025-10-04 04:29:56.713558182 +0000 UTC m=+5208.005567894" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.154517 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7"] Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.157737 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.160604 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.164630 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.189810 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7"] Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.271067 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-qhmd4"] Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.272300 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qhmd4" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.276098 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qhmd4"] Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.334696 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51cff2d-a9f1-4262-9592-c4fc36dc299f-config-volume\") pod \"collect-profiles-29325870-6j6t7\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.334747 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51cff2d-a9f1-4262-9592-c4fc36dc299f-secret-volume\") pod \"collect-profiles-29325870-6j6t7\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.335092 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkx5g\" (UniqueName: \"kubernetes.io/projected/b51cff2d-a9f1-4262-9592-c4fc36dc299f-kube-api-access-vkx5g\") pod \"collect-profiles-29325870-6j6t7\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.436928 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqf58\" (UniqueName: \"kubernetes.io/projected/147d8093-1179-48bb-a771-8cb1b6a7f8a7-kube-api-access-zqf58\") pod \"keystone-db-create-qhmd4\" (UID: \"147d8093-1179-48bb-a771-8cb1b6a7f8a7\") " pod="openstack/keystone-db-create-qhmd4" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.437078 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkx5g\" (UniqueName: \"kubernetes.io/projected/b51cff2d-a9f1-4262-9592-c4fc36dc299f-kube-api-access-vkx5g\") pod \"collect-profiles-29325870-6j6t7\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.437145 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51cff2d-a9f1-4262-9592-c4fc36dc299f-config-volume\") pod \"collect-profiles-29325870-6j6t7\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.437182 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51cff2d-a9f1-4262-9592-c4fc36dc299f-secret-volume\") pod \"collect-profiles-29325870-6j6t7\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.438545 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51cff2d-a9f1-4262-9592-c4fc36dc299f-config-volume\") pod \"collect-profiles-29325870-6j6t7\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.444328 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51cff2d-a9f1-4262-9592-c4fc36dc299f-secret-volume\") pod \"collect-profiles-29325870-6j6t7\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.457384 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkx5g\" (UniqueName: \"kubernetes.io/projected/b51cff2d-a9f1-4262-9592-c4fc36dc299f-kube-api-access-vkx5g\") pod \"collect-profiles-29325870-6j6t7\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.489411 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.538982 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqf58\" (UniqueName: \"kubernetes.io/projected/147d8093-1179-48bb-a771-8cb1b6a7f8a7-kube-api-access-zqf58\") pod \"keystone-db-create-qhmd4\" (UID: \"147d8093-1179-48bb-a771-8cb1b6a7f8a7\") " pod="openstack/keystone-db-create-qhmd4" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.560674 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqf58\" (UniqueName: \"kubernetes.io/projected/147d8093-1179-48bb-a771-8cb1b6a7f8a7-kube-api-access-zqf58\") pod \"keystone-db-create-qhmd4\" (UID: \"147d8093-1179-48bb-a771-8cb1b6a7f8a7\") " pod="openstack/keystone-db-create-qhmd4" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.595416 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qhmd4" Oct 04 04:30:00 crc kubenswrapper[4770]: I1004 04:30:00.929865 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7"] Oct 04 04:30:00 crc kubenswrapper[4770]: W1004 04:30:00.933058 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb51cff2d_a9f1_4262_9592_c4fc36dc299f.slice/crio-5772fab3ae983287367b10065118c6f94ce3467aa83a75711378ae896102b1f1 WatchSource:0}: Error finding container 5772fab3ae983287367b10065118c6f94ce3467aa83a75711378ae896102b1f1: Status 404 returned error can't find the container with id 5772fab3ae983287367b10065118c6f94ce3467aa83a75711378ae896102b1f1 Oct 04 04:30:01 crc kubenswrapper[4770]: W1004 04:30:01.051663 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod147d8093_1179_48bb_a771_8cb1b6a7f8a7.slice/crio-7722d2a8604d9e7ce2ad114c0b621a7afbc772626f956c2ea278ff8ea2744ece WatchSource:0}: Error finding container 7722d2a8604d9e7ce2ad114c0b621a7afbc772626f956c2ea278ff8ea2744ece: Status 404 returned error can't find the container with id 7722d2a8604d9e7ce2ad114c0b621a7afbc772626f956c2ea278ff8ea2744ece Oct 04 04:30:01 crc kubenswrapper[4770]: I1004 04:30:01.052293 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qhmd4"] Oct 04 04:30:01 crc kubenswrapper[4770]: I1004 04:30:01.752941 4770 generic.go:334] "Generic (PLEG): container finished" podID="147d8093-1179-48bb-a771-8cb1b6a7f8a7" containerID="1312cd478db70cdc0313cd3744c6a773b954d9be4c176178827f9ca822adf4bc" exitCode=0 Oct 04 04:30:01 crc kubenswrapper[4770]: I1004 04:30:01.752992 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qhmd4" event={"ID":"147d8093-1179-48bb-a771-8cb1b6a7f8a7","Type":"ContainerDied","Data":"1312cd478db70cdc0313cd3744c6a773b954d9be4c176178827f9ca822adf4bc"} Oct 04 04:30:01 crc kubenswrapper[4770]: I1004 04:30:01.753330 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qhmd4" event={"ID":"147d8093-1179-48bb-a771-8cb1b6a7f8a7","Type":"ContainerStarted","Data":"7722d2a8604d9e7ce2ad114c0b621a7afbc772626f956c2ea278ff8ea2744ece"} Oct 04 04:30:01 crc kubenswrapper[4770]: I1004 04:30:01.756721 4770 generic.go:334] "Generic (PLEG): container finished" podID="b51cff2d-a9f1-4262-9592-c4fc36dc299f" containerID="cd6c1de9e081ea52546c1269f24f0471edf2fcd487697b6a21c172d5a68941c5" exitCode=0 Oct 04 04:30:01 crc kubenswrapper[4770]: I1004 04:30:01.756782 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" event={"ID":"b51cff2d-a9f1-4262-9592-c4fc36dc299f","Type":"ContainerDied","Data":"cd6c1de9e081ea52546c1269f24f0471edf2fcd487697b6a21c172d5a68941c5"} Oct 04 04:30:01 crc kubenswrapper[4770]: I1004 04:30:01.756818 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" event={"ID":"b51cff2d-a9f1-4262-9592-c4fc36dc299f","Type":"ContainerStarted","Data":"5772fab3ae983287367b10065118c6f94ce3467aa83a75711378ae896102b1f1"} Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.169905 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.180058 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qhmd4" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.290228 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkx5g\" (UniqueName: \"kubernetes.io/projected/b51cff2d-a9f1-4262-9592-c4fc36dc299f-kube-api-access-vkx5g\") pod \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.290270 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51cff2d-a9f1-4262-9592-c4fc36dc299f-config-volume\") pod \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.290300 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51cff2d-a9f1-4262-9592-c4fc36dc299f-secret-volume\") pod \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\" (UID: \"b51cff2d-a9f1-4262-9592-c4fc36dc299f\") " Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.290435 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqf58\" (UniqueName: \"kubernetes.io/projected/147d8093-1179-48bb-a771-8cb1b6a7f8a7-kube-api-access-zqf58\") pod \"147d8093-1179-48bb-a771-8cb1b6a7f8a7\" (UID: \"147d8093-1179-48bb-a771-8cb1b6a7f8a7\") " Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.290899 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b51cff2d-a9f1-4262-9592-c4fc36dc299f-config-volume" (OuterVolumeSpecName: "config-volume") pod "b51cff2d-a9f1-4262-9592-c4fc36dc299f" (UID: "b51cff2d-a9f1-4262-9592-c4fc36dc299f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.296198 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51cff2d-a9f1-4262-9592-c4fc36dc299f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b51cff2d-a9f1-4262-9592-c4fc36dc299f" (UID: "b51cff2d-a9f1-4262-9592-c4fc36dc299f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.311050 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51cff2d-a9f1-4262-9592-c4fc36dc299f-kube-api-access-vkx5g" (OuterVolumeSpecName: "kube-api-access-vkx5g") pod "b51cff2d-a9f1-4262-9592-c4fc36dc299f" (UID: "b51cff2d-a9f1-4262-9592-c4fc36dc299f"). InnerVolumeSpecName "kube-api-access-vkx5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.311131 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147d8093-1179-48bb-a771-8cb1b6a7f8a7-kube-api-access-zqf58" (OuterVolumeSpecName: "kube-api-access-zqf58") pod "147d8093-1179-48bb-a771-8cb1b6a7f8a7" (UID: "147d8093-1179-48bb-a771-8cb1b6a7f8a7"). InnerVolumeSpecName "kube-api-access-zqf58". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.392129 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkx5g\" (UniqueName: \"kubernetes.io/projected/b51cff2d-a9f1-4262-9592-c4fc36dc299f-kube-api-access-vkx5g\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.392413 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b51cff2d-a9f1-4262-9592-c4fc36dc299f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.392491 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b51cff2d-a9f1-4262-9592-c4fc36dc299f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.392594 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqf58\" (UniqueName: \"kubernetes.io/projected/147d8093-1179-48bb-a771-8cb1b6a7f8a7-kube-api-access-zqf58\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.777766 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" event={"ID":"b51cff2d-a9f1-4262-9592-c4fc36dc299f","Type":"ContainerDied","Data":"5772fab3ae983287367b10065118c6f94ce3467aa83a75711378ae896102b1f1"} Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.777823 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5772fab3ae983287367b10065118c6f94ce3467aa83a75711378ae896102b1f1" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.777774 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.780053 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qhmd4" event={"ID":"147d8093-1179-48bb-a771-8cb1b6a7f8a7","Type":"ContainerDied","Data":"7722d2a8604d9e7ce2ad114c0b621a7afbc772626f956c2ea278ff8ea2744ece"} Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.780088 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qhmd4" Oct 04 04:30:03 crc kubenswrapper[4770]: I1004 04:30:03.780091 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7722d2a8604d9e7ce2ad114c0b621a7afbc772626f956c2ea278ff8ea2744ece" Oct 04 04:30:04 crc kubenswrapper[4770]: I1004 04:30:04.263434 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2"] Oct 04 04:30:04 crc kubenswrapper[4770]: I1004 04:30:04.272925 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-4chq2"] Oct 04 04:30:05 crc kubenswrapper[4770]: I1004 04:30:05.690601 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e497921b-5585-45d0-97a1-1330033db38e" path="/var/lib/kubelet/pods/e497921b-5585-45d0-97a1-1330033db38e/volumes" Oct 04 04:30:09 crc kubenswrapper[4770]: I1004 04:30:09.754503 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.398812 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-44b0-account-create-cpdz8"] Oct 04 04:30:10 crc kubenswrapper[4770]: E1004 04:30:10.399508 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51cff2d-a9f1-4262-9592-c4fc36dc299f" containerName="collect-profiles" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.399529 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51cff2d-a9f1-4262-9592-c4fc36dc299f" containerName="collect-profiles" Oct 04 04:30:10 crc kubenswrapper[4770]: E1004 04:30:10.399577 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147d8093-1179-48bb-a771-8cb1b6a7f8a7" containerName="mariadb-database-create" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.399589 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="147d8093-1179-48bb-a771-8cb1b6a7f8a7" containerName="mariadb-database-create" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.399774 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51cff2d-a9f1-4262-9592-c4fc36dc299f" containerName="collect-profiles" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.399806 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="147d8093-1179-48bb-a771-8cb1b6a7f8a7" containerName="mariadb-database-create" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.400584 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-44b0-account-create-cpdz8" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.402398 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.413698 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-44b0-account-create-cpdz8"] Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.513345 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79gp6\" (UniqueName: \"kubernetes.io/projected/a7101c35-8518-43e9-8a1f-fed84491a6dc-kube-api-access-79gp6\") pod \"keystone-44b0-account-create-cpdz8\" (UID: \"a7101c35-8518-43e9-8a1f-fed84491a6dc\") " pod="openstack/keystone-44b0-account-create-cpdz8" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.615093 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79gp6\" (UniqueName: \"kubernetes.io/projected/a7101c35-8518-43e9-8a1f-fed84491a6dc-kube-api-access-79gp6\") pod \"keystone-44b0-account-create-cpdz8\" (UID: \"a7101c35-8518-43e9-8a1f-fed84491a6dc\") " pod="openstack/keystone-44b0-account-create-cpdz8" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.635584 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79gp6\" (UniqueName: \"kubernetes.io/projected/a7101c35-8518-43e9-8a1f-fed84491a6dc-kube-api-access-79gp6\") pod \"keystone-44b0-account-create-cpdz8\" (UID: \"a7101c35-8518-43e9-8a1f-fed84491a6dc\") " pod="openstack/keystone-44b0-account-create-cpdz8" Oct 04 04:30:10 crc kubenswrapper[4770]: I1004 04:30:10.726154 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-44b0-account-create-cpdz8" Oct 04 04:30:11 crc kubenswrapper[4770]: I1004 04:30:11.199502 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-44b0-account-create-cpdz8"] Oct 04 04:30:11 crc kubenswrapper[4770]: I1004 04:30:11.890851 4770 generic.go:334] "Generic (PLEG): container finished" podID="a7101c35-8518-43e9-8a1f-fed84491a6dc" containerID="bb45933b28e6a908156865440b73265b4f3d6c8febdcece1962dcda742904783" exitCode=0 Oct 04 04:30:11 crc kubenswrapper[4770]: I1004 04:30:11.891338 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-44b0-account-create-cpdz8" event={"ID":"a7101c35-8518-43e9-8a1f-fed84491a6dc","Type":"ContainerDied","Data":"bb45933b28e6a908156865440b73265b4f3d6c8febdcece1962dcda742904783"} Oct 04 04:30:11 crc kubenswrapper[4770]: I1004 04:30:11.891382 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-44b0-account-create-cpdz8" event={"ID":"a7101c35-8518-43e9-8a1f-fed84491a6dc","Type":"ContainerStarted","Data":"6f4d1c04e5d38f8139b6ec5224ef2a17acc0d1599d1f90f0d8ab1d957871cff8"} Oct 04 04:30:13 crc kubenswrapper[4770]: I1004 04:30:13.192665 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-44b0-account-create-cpdz8" Oct 04 04:30:13 crc kubenswrapper[4770]: I1004 04:30:13.261550 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79gp6\" (UniqueName: \"kubernetes.io/projected/a7101c35-8518-43e9-8a1f-fed84491a6dc-kube-api-access-79gp6\") pod \"a7101c35-8518-43e9-8a1f-fed84491a6dc\" (UID: \"a7101c35-8518-43e9-8a1f-fed84491a6dc\") " Oct 04 04:30:13 crc kubenswrapper[4770]: I1004 04:30:13.274777 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7101c35-8518-43e9-8a1f-fed84491a6dc-kube-api-access-79gp6" (OuterVolumeSpecName: "kube-api-access-79gp6") pod "a7101c35-8518-43e9-8a1f-fed84491a6dc" (UID: "a7101c35-8518-43e9-8a1f-fed84491a6dc"). InnerVolumeSpecName "kube-api-access-79gp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:30:13 crc kubenswrapper[4770]: I1004 04:30:13.374799 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79gp6\" (UniqueName: \"kubernetes.io/projected/a7101c35-8518-43e9-8a1f-fed84491a6dc-kube-api-access-79gp6\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:13 crc kubenswrapper[4770]: I1004 04:30:13.907690 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-44b0-account-create-cpdz8" event={"ID":"a7101c35-8518-43e9-8a1f-fed84491a6dc","Type":"ContainerDied","Data":"6f4d1c04e5d38f8139b6ec5224ef2a17acc0d1599d1f90f0d8ab1d957871cff8"} Oct 04 04:30:13 crc kubenswrapper[4770]: I1004 04:30:13.907747 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f4d1c04e5d38f8139b6ec5224ef2a17acc0d1599d1f90f0d8ab1d957871cff8" Oct 04 04:30:13 crc kubenswrapper[4770]: I1004 04:30:13.907747 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-44b0-account-create-cpdz8" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.810248 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-dbmp8"] Oct 04 04:30:15 crc kubenswrapper[4770]: E1004 04:30:15.811023 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7101c35-8518-43e9-8a1f-fed84491a6dc" containerName="mariadb-account-create" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.811039 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7101c35-8518-43e9-8a1f-fed84491a6dc" containerName="mariadb-account-create" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.811228 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7101c35-8518-43e9-8a1f-fed84491a6dc" containerName="mariadb-account-create" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.811901 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.813922 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-whxb7" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.813934 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.814313 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.814727 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.821870 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dbmp8"] Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.916448 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-config-data\") pod \"keystone-db-sync-dbmp8\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.917004 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8qt4\" (UniqueName: \"kubernetes.io/projected/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-kube-api-access-w8qt4\") pod \"keystone-db-sync-dbmp8\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:15 crc kubenswrapper[4770]: I1004 04:30:15.917103 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-combined-ca-bundle\") pod \"keystone-db-sync-dbmp8\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.018416 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-config-data\") pod \"keystone-db-sync-dbmp8\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.018581 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8qt4\" (UniqueName: \"kubernetes.io/projected/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-kube-api-access-w8qt4\") pod \"keystone-db-sync-dbmp8\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.018619 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-combined-ca-bundle\") pod \"keystone-db-sync-dbmp8\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.025805 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-combined-ca-bundle\") pod \"keystone-db-sync-dbmp8\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.026351 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-config-data\") pod \"keystone-db-sync-dbmp8\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.037664 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8qt4\" (UniqueName: \"kubernetes.io/projected/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-kube-api-access-w8qt4\") pod \"keystone-db-sync-dbmp8\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.129617 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.631215 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dbmp8"] Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.933605 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dbmp8" event={"ID":"c3274a0e-8276-49f0-9583-ef9bfdb43c6e","Type":"ContainerStarted","Data":"06f2ed82f10c96ad2ef5cf70956895ed9d0d0c77a294c20c2ed0adb7c6f8dfa9"} Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.934061 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dbmp8" event={"ID":"c3274a0e-8276-49f0-9583-ef9bfdb43c6e","Type":"ContainerStarted","Data":"8d7fc5878c6e351d21cdfb19a117ced793e810413a4e9250dc5073c6293707ec"} Oct 04 04:30:16 crc kubenswrapper[4770]: I1004 04:30:16.970501 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-dbmp8" podStartSLOduration=1.97047406 podStartE2EDuration="1.97047406s" podCreationTimestamp="2025-10-04 04:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:30:16.962102651 +0000 UTC m=+5228.254112423" watchObservedRunningTime="2025-10-04 04:30:16.97047406 +0000 UTC m=+5228.262483802" Oct 04 04:30:18 crc kubenswrapper[4770]: I1004 04:30:18.964987 4770 generic.go:334] "Generic (PLEG): container finished" podID="c3274a0e-8276-49f0-9583-ef9bfdb43c6e" containerID="06f2ed82f10c96ad2ef5cf70956895ed9d0d0c77a294c20c2ed0adb7c6f8dfa9" exitCode=0 Oct 04 04:30:18 crc kubenswrapper[4770]: I1004 04:30:18.965093 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dbmp8" event={"ID":"c3274a0e-8276-49f0-9583-ef9bfdb43c6e","Type":"ContainerDied","Data":"06f2ed82f10c96ad2ef5cf70956895ed9d0d0c77a294c20c2ed0adb7c6f8dfa9"} Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.266386 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.396792 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8qt4\" (UniqueName: \"kubernetes.io/projected/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-kube-api-access-w8qt4\") pod \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.396965 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-combined-ca-bundle\") pod \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.397088 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-config-data\") pod \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\" (UID: \"c3274a0e-8276-49f0-9583-ef9bfdb43c6e\") " Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.402626 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-kube-api-access-w8qt4" (OuterVolumeSpecName: "kube-api-access-w8qt4") pod "c3274a0e-8276-49f0-9583-ef9bfdb43c6e" (UID: "c3274a0e-8276-49f0-9583-ef9bfdb43c6e"). InnerVolumeSpecName "kube-api-access-w8qt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.424841 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3274a0e-8276-49f0-9583-ef9bfdb43c6e" (UID: "c3274a0e-8276-49f0-9583-ef9bfdb43c6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.451258 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-config-data" (OuterVolumeSpecName: "config-data") pod "c3274a0e-8276-49f0-9583-ef9bfdb43c6e" (UID: "c3274a0e-8276-49f0-9583-ef9bfdb43c6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.498391 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.498424 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.498435 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8qt4\" (UniqueName: \"kubernetes.io/projected/c3274a0e-8276-49f0-9583-ef9bfdb43c6e-kube-api-access-w8qt4\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.986539 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dbmp8" event={"ID":"c3274a0e-8276-49f0-9583-ef9bfdb43c6e","Type":"ContainerDied","Data":"8d7fc5878c6e351d21cdfb19a117ced793e810413a4e9250dc5073c6293707ec"} Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.986598 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d7fc5878c6e351d21cdfb19a117ced793e810413a4e9250dc5073c6293707ec" Oct 04 04:30:20 crc kubenswrapper[4770]: I1004 04:30:20.986648 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dbmp8" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.250678 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59f4ccb565-qspsj"] Oct 04 04:30:21 crc kubenswrapper[4770]: E1004 04:30:21.251140 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3274a0e-8276-49f0-9583-ef9bfdb43c6e" containerName="keystone-db-sync" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.251159 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3274a0e-8276-49f0-9583-ef9bfdb43c6e" containerName="keystone-db-sync" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.260306 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3274a0e-8276-49f0-9583-ef9bfdb43c6e" containerName="keystone-db-sync" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.280302 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59f4ccb565-qspsj"] Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.280428 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.307130 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2sz6v"] Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.309851 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.315887 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.316033 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-whxb7" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.316071 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2sz6v"] Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.315906 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.316139 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.418501 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-sb\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.418565 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-dns-svc\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.418618 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4g8r\" (UniqueName: \"kubernetes.io/projected/2e0915a5-e705-43ca-962e-8dde99e19499-kube-api-access-w4g8r\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.418831 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-credential-keys\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.418938 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9jfc\" (UniqueName: \"kubernetes.io/projected/cb075f34-9b51-467d-978c-36c563b48d81-kube-api-access-f9jfc\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.418967 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-scripts\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.419057 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-fernet-keys\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.419127 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-config-data\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.419215 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-nb\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.419280 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-combined-ca-bundle\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.419311 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-config\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.521875 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-config-data\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.522528 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-nb\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.522614 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-combined-ca-bundle\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.522670 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-config\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.522768 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-sb\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.522834 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-dns-svc\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.522931 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4g8r\" (UniqueName: \"kubernetes.io/projected/2e0915a5-e705-43ca-962e-8dde99e19499-kube-api-access-w4g8r\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.523083 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-credential-keys\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.524180 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-dns-svc\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.524485 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-nb\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.524269 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-sb\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.524705 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9jfc\" (UniqueName: \"kubernetes.io/projected/cb075f34-9b51-467d-978c-36c563b48d81-kube-api-access-f9jfc\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.524741 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-scripts\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.524807 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-fernet-keys\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.524827 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-config\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.530511 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-combined-ca-bundle\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.530781 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-config-data\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.531765 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-scripts\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.533779 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-credential-keys\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.536986 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-fernet-keys\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.544989 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9jfc\" (UniqueName: \"kubernetes.io/projected/cb075f34-9b51-467d-978c-36c563b48d81-kube-api-access-f9jfc\") pod \"dnsmasq-dns-59f4ccb565-qspsj\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.546433 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4g8r\" (UniqueName: \"kubernetes.io/projected/2e0915a5-e705-43ca-962e-8dde99e19499-kube-api-access-w4g8r\") pod \"keystone-bootstrap-2sz6v\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.606989 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.627852 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.965426 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2sz6v"] Oct 04 04:30:21 crc kubenswrapper[4770]: W1004 04:30:21.971896 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e0915a5_e705_43ca_962e_8dde99e19499.slice/crio-5745664ba11854f813eb5b8389095e1eb1f3c9054cbee124daba7beb57eb9da8 WatchSource:0}: Error finding container 5745664ba11854f813eb5b8389095e1eb1f3c9054cbee124daba7beb57eb9da8: Status 404 returned error can't find the container with id 5745664ba11854f813eb5b8389095e1eb1f3c9054cbee124daba7beb57eb9da8 Oct 04 04:30:21 crc kubenswrapper[4770]: I1004 04:30:21.996892 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2sz6v" event={"ID":"2e0915a5-e705-43ca-962e-8dde99e19499","Type":"ContainerStarted","Data":"5745664ba11854f813eb5b8389095e1eb1f3c9054cbee124daba7beb57eb9da8"} Oct 04 04:30:22 crc kubenswrapper[4770]: I1004 04:30:22.086595 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59f4ccb565-qspsj"] Oct 04 04:30:22 crc kubenswrapper[4770]: W1004 04:30:22.095113 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb075f34_9b51_467d_978c_36c563b48d81.slice/crio-3375981a47cee8018ce9760a8e3aeb469c681f534772ebc953e6b1f92641640f WatchSource:0}: Error finding container 3375981a47cee8018ce9760a8e3aeb469c681f534772ebc953e6b1f92641640f: Status 404 returned error can't find the container with id 3375981a47cee8018ce9760a8e3aeb469c681f534772ebc953e6b1f92641640f Oct 04 04:30:23 crc kubenswrapper[4770]: I1004 04:30:23.014746 4770 generic.go:334] "Generic (PLEG): container finished" podID="cb075f34-9b51-467d-978c-36c563b48d81" containerID="0f218d03b43bb2375c2357953820aa65de71bb15fa5eab443fdacd6a6f4a6e39" exitCode=0 Oct 04 04:30:23 crc kubenswrapper[4770]: I1004 04:30:23.014825 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" event={"ID":"cb075f34-9b51-467d-978c-36c563b48d81","Type":"ContainerDied","Data":"0f218d03b43bb2375c2357953820aa65de71bb15fa5eab443fdacd6a6f4a6e39"} Oct 04 04:30:23 crc kubenswrapper[4770]: I1004 04:30:23.015318 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" event={"ID":"cb075f34-9b51-467d-978c-36c563b48d81","Type":"ContainerStarted","Data":"3375981a47cee8018ce9760a8e3aeb469c681f534772ebc953e6b1f92641640f"} Oct 04 04:30:23 crc kubenswrapper[4770]: I1004 04:30:23.024363 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2sz6v" event={"ID":"2e0915a5-e705-43ca-962e-8dde99e19499","Type":"ContainerStarted","Data":"6cb4056e8832a1959563883de74266a80f382bfddd48a1b55ccb6a0ea3a8481d"} Oct 04 04:30:23 crc kubenswrapper[4770]: I1004 04:30:23.082443 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2sz6v" podStartSLOduration=2.082400051 podStartE2EDuration="2.082400051s" podCreationTimestamp="2025-10-04 04:30:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:30:23.074250247 +0000 UTC m=+5234.366259979" watchObservedRunningTime="2025-10-04 04:30:23.082400051 +0000 UTC m=+5234.374409773" Oct 04 04:30:24 crc kubenswrapper[4770]: I1004 04:30:24.035925 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" event={"ID":"cb075f34-9b51-467d-978c-36c563b48d81","Type":"ContainerStarted","Data":"116bce903fb7391ca951e960857439b3fbff228a9dcc1a3956168bba9e9f8871"} Oct 04 04:30:24 crc kubenswrapper[4770]: I1004 04:30:24.036301 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:24 crc kubenswrapper[4770]: I1004 04:30:24.058764 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" podStartSLOduration=3.058737619 podStartE2EDuration="3.058737619s" podCreationTimestamp="2025-10-04 04:30:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:30:24.055992507 +0000 UTC m=+5235.348002259" watchObservedRunningTime="2025-10-04 04:30:24.058737619 +0000 UTC m=+5235.350747351" Oct 04 04:30:24 crc kubenswrapper[4770]: I1004 04:30:24.786577 4770 scope.go:117] "RemoveContainer" containerID="5757b45d9ce0e40de8c5a1cb226b51a2cc8b505db4064e90846bd02386b7180b" Oct 04 04:30:24 crc kubenswrapper[4770]: I1004 04:30:24.812441 4770 scope.go:117] "RemoveContainer" containerID="108db819f7b370469975ad54fd596e8fbd090999a05ccdde7714e0f7b2026b04" Oct 04 04:30:26 crc kubenswrapper[4770]: I1004 04:30:26.066291 4770 generic.go:334] "Generic (PLEG): container finished" podID="2e0915a5-e705-43ca-962e-8dde99e19499" containerID="6cb4056e8832a1959563883de74266a80f382bfddd48a1b55ccb6a0ea3a8481d" exitCode=0 Oct 04 04:30:26 crc kubenswrapper[4770]: I1004 04:30:26.066359 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2sz6v" event={"ID":"2e0915a5-e705-43ca-962e-8dde99e19499","Type":"ContainerDied","Data":"6cb4056e8832a1959563883de74266a80f382bfddd48a1b55ccb6a0ea3a8481d"} Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.424674 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.563496 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4g8r\" (UniqueName: \"kubernetes.io/projected/2e0915a5-e705-43ca-962e-8dde99e19499-kube-api-access-w4g8r\") pod \"2e0915a5-e705-43ca-962e-8dde99e19499\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.563570 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-fernet-keys\") pod \"2e0915a5-e705-43ca-962e-8dde99e19499\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.563610 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-config-data\") pod \"2e0915a5-e705-43ca-962e-8dde99e19499\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.563753 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-credential-keys\") pod \"2e0915a5-e705-43ca-962e-8dde99e19499\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.563807 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-scripts\") pod \"2e0915a5-e705-43ca-962e-8dde99e19499\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.563894 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-combined-ca-bundle\") pod \"2e0915a5-e705-43ca-962e-8dde99e19499\" (UID: \"2e0915a5-e705-43ca-962e-8dde99e19499\") " Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.570103 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2e0915a5-e705-43ca-962e-8dde99e19499" (UID: "2e0915a5-e705-43ca-962e-8dde99e19499"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.570524 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-scripts" (OuterVolumeSpecName: "scripts") pod "2e0915a5-e705-43ca-962e-8dde99e19499" (UID: "2e0915a5-e705-43ca-962e-8dde99e19499"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.571093 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e0915a5-e705-43ca-962e-8dde99e19499-kube-api-access-w4g8r" (OuterVolumeSpecName: "kube-api-access-w4g8r") pod "2e0915a5-e705-43ca-962e-8dde99e19499" (UID: "2e0915a5-e705-43ca-962e-8dde99e19499"). InnerVolumeSpecName "kube-api-access-w4g8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.571645 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2e0915a5-e705-43ca-962e-8dde99e19499" (UID: "2e0915a5-e705-43ca-962e-8dde99e19499"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.588441 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e0915a5-e705-43ca-962e-8dde99e19499" (UID: "2e0915a5-e705-43ca-962e-8dde99e19499"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.592545 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-config-data" (OuterVolumeSpecName: "config-data") pod "2e0915a5-e705-43ca-962e-8dde99e19499" (UID: "2e0915a5-e705-43ca-962e-8dde99e19499"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.666730 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4g8r\" (UniqueName: \"kubernetes.io/projected/2e0915a5-e705-43ca-962e-8dde99e19499-kube-api-access-w4g8r\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.667061 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.667097 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.667120 4770 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.667143 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:27 crc kubenswrapper[4770]: I1004 04:30:27.667167 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0915a5-e705-43ca-962e-8dde99e19499-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.090562 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2sz6v" event={"ID":"2e0915a5-e705-43ca-962e-8dde99e19499","Type":"ContainerDied","Data":"5745664ba11854f813eb5b8389095e1eb1f3c9054cbee124daba7beb57eb9da8"} Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.090607 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5745664ba11854f813eb5b8389095e1eb1f3c9054cbee124daba7beb57eb9da8" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.090659 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2sz6v" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.300560 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2sz6v"] Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.309524 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2sz6v"] Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.382746 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fkn6b"] Oct 04 04:30:28 crc kubenswrapper[4770]: E1004 04:30:28.383155 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e0915a5-e705-43ca-962e-8dde99e19499" containerName="keystone-bootstrap" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.383170 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e0915a5-e705-43ca-962e-8dde99e19499" containerName="keystone-bootstrap" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.383363 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e0915a5-e705-43ca-962e-8dde99e19499" containerName="keystone-bootstrap" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.383943 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.387187 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.387215 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.387225 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-whxb7" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.390942 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.394753 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fkn6b"] Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.480348 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-scripts\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.480407 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-config-data\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.480490 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-fernet-keys\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.480624 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-combined-ca-bundle\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.480697 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-credential-keys\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.480858 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm24r\" (UniqueName: \"kubernetes.io/projected/a3706ecc-dda1-498f-8319-b10379ceccb9-kube-api-access-tm24r\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.582182 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm24r\" (UniqueName: \"kubernetes.io/projected/a3706ecc-dda1-498f-8319-b10379ceccb9-kube-api-access-tm24r\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.582259 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-scripts\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.582288 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-config-data\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.582327 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-fernet-keys\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.582356 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-combined-ca-bundle\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.582376 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-credential-keys\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.586091 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-config-data\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.586277 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-scripts\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.586295 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-combined-ca-bundle\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.586551 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-fernet-keys\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.587144 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-credential-keys\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.603405 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm24r\" (UniqueName: \"kubernetes.io/projected/a3706ecc-dda1-498f-8319-b10379ceccb9-kube-api-access-tm24r\") pod \"keystone-bootstrap-fkn6b\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:28 crc kubenswrapper[4770]: I1004 04:30:28.704180 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:29 crc kubenswrapper[4770]: I1004 04:30:29.218441 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fkn6b"] Oct 04 04:30:29 crc kubenswrapper[4770]: I1004 04:30:29.695747 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e0915a5-e705-43ca-962e-8dde99e19499" path="/var/lib/kubelet/pods/2e0915a5-e705-43ca-962e-8dde99e19499/volumes" Oct 04 04:30:30 crc kubenswrapper[4770]: I1004 04:30:30.116245 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fkn6b" event={"ID":"a3706ecc-dda1-498f-8319-b10379ceccb9","Type":"ContainerStarted","Data":"f8ac077e3dcd2450d464e3c24e55a86c463b22fe37fd2483ae6a96cfc74f11a0"} Oct 04 04:30:30 crc kubenswrapper[4770]: I1004 04:30:30.116289 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fkn6b" event={"ID":"a3706ecc-dda1-498f-8319-b10379ceccb9","Type":"ContainerStarted","Data":"cfcb5dd7af4a356e567d04c0315d5489de9586ba3e24bc05394d11b66d698ed8"} Oct 04 04:30:30 crc kubenswrapper[4770]: I1004 04:30:30.157584 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fkn6b" podStartSLOduration=2.157555687 podStartE2EDuration="2.157555687s" podCreationTimestamp="2025-10-04 04:30:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:30:30.14847347 +0000 UTC m=+5241.440483212" watchObservedRunningTime="2025-10-04 04:30:30.157555687 +0000 UTC m=+5241.449565439" Oct 04 04:30:31 crc kubenswrapper[4770]: I1004 04:30:31.608896 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:30:31 crc kubenswrapper[4770]: I1004 04:30:31.656104 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778dbf57b9-d5qp2"] Oct 04 04:30:31 crc kubenswrapper[4770]: I1004 04:30:31.656364 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" podUID="381071ee-0780-4ba9-a924-4da9ea4113b2" containerName="dnsmasq-dns" containerID="cri-o://958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a" gracePeriod=10 Oct 04 04:30:31 crc kubenswrapper[4770]: I1004 04:30:31.795152 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:30:31 crc kubenswrapper[4770]: I1004 04:30:31.795206 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.111740 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.141416 4770 generic.go:334] "Generic (PLEG): container finished" podID="381071ee-0780-4ba9-a924-4da9ea4113b2" containerID="958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a" exitCode=0 Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.141462 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" event={"ID":"381071ee-0780-4ba9-a924-4da9ea4113b2","Type":"ContainerDied","Data":"958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a"} Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.141497 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" event={"ID":"381071ee-0780-4ba9-a924-4da9ea4113b2","Type":"ContainerDied","Data":"57a1f4d7720ef139005e93af488ea6be11f74b5666da63fd8889069313ebc829"} Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.141518 4770 scope.go:117] "RemoveContainer" containerID="958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.141910 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778dbf57b9-d5qp2" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.165472 4770 scope.go:117] "RemoveContainer" containerID="123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.198978 4770 scope.go:117] "RemoveContainer" containerID="958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a" Oct 04 04:30:32 crc kubenswrapper[4770]: E1004 04:30:32.199632 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a\": container with ID starting with 958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a not found: ID does not exist" containerID="958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.199688 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a"} err="failed to get container status \"958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a\": rpc error: code = NotFound desc = could not find container \"958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a\": container with ID starting with 958d40d1866680848a4f85cf3b399f0d79681770f345a5040a50255b0be8056a not found: ID does not exist" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.199761 4770 scope.go:117] "RemoveContainer" containerID="123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86" Oct 04 04:30:32 crc kubenswrapper[4770]: E1004 04:30:32.200222 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86\": container with ID starting with 123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86 not found: ID does not exist" containerID="123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.200257 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86"} err="failed to get container status \"123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86\": rpc error: code = NotFound desc = could not find container \"123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86\": container with ID starting with 123ad988c67c40084b4d88a7bdf48088dde0ce0f6bee4a7b93caf1d19fed4f86 not found: ID does not exist" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.252548 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-sb\") pod \"381071ee-0780-4ba9-a924-4da9ea4113b2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.252736 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twz8r\" (UniqueName: \"kubernetes.io/projected/381071ee-0780-4ba9-a924-4da9ea4113b2-kube-api-access-twz8r\") pod \"381071ee-0780-4ba9-a924-4da9ea4113b2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.252758 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-config\") pod \"381071ee-0780-4ba9-a924-4da9ea4113b2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.252789 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-nb\") pod \"381071ee-0780-4ba9-a924-4da9ea4113b2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.252813 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-dns-svc\") pod \"381071ee-0780-4ba9-a924-4da9ea4113b2\" (UID: \"381071ee-0780-4ba9-a924-4da9ea4113b2\") " Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.262597 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/381071ee-0780-4ba9-a924-4da9ea4113b2-kube-api-access-twz8r" (OuterVolumeSpecName: "kube-api-access-twz8r") pod "381071ee-0780-4ba9-a924-4da9ea4113b2" (UID: "381071ee-0780-4ba9-a924-4da9ea4113b2"). InnerVolumeSpecName "kube-api-access-twz8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.327596 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "381071ee-0780-4ba9-a924-4da9ea4113b2" (UID: "381071ee-0780-4ba9-a924-4da9ea4113b2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.337481 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "381071ee-0780-4ba9-a924-4da9ea4113b2" (UID: "381071ee-0780-4ba9-a924-4da9ea4113b2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.354096 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twz8r\" (UniqueName: \"kubernetes.io/projected/381071ee-0780-4ba9-a924-4da9ea4113b2-kube-api-access-twz8r\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.354362 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.354461 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.365766 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-config" (OuterVolumeSpecName: "config") pod "381071ee-0780-4ba9-a924-4da9ea4113b2" (UID: "381071ee-0780-4ba9-a924-4da9ea4113b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.371854 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "381071ee-0780-4ba9-a924-4da9ea4113b2" (UID: "381071ee-0780-4ba9-a924-4da9ea4113b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.456502 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.456637 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/381071ee-0780-4ba9-a924-4da9ea4113b2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.474378 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778dbf57b9-d5qp2"] Oct 04 04:30:32 crc kubenswrapper[4770]: I1004 04:30:32.481002 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-778dbf57b9-d5qp2"] Oct 04 04:30:33 crc kubenswrapper[4770]: I1004 04:30:33.162993 4770 generic.go:334] "Generic (PLEG): container finished" podID="a3706ecc-dda1-498f-8319-b10379ceccb9" containerID="f8ac077e3dcd2450d464e3c24e55a86c463b22fe37fd2483ae6a96cfc74f11a0" exitCode=0 Oct 04 04:30:33 crc kubenswrapper[4770]: I1004 04:30:33.163295 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fkn6b" event={"ID":"a3706ecc-dda1-498f-8319-b10379ceccb9","Type":"ContainerDied","Data":"f8ac077e3dcd2450d464e3c24e55a86c463b22fe37fd2483ae6a96cfc74f11a0"} Oct 04 04:30:33 crc kubenswrapper[4770]: I1004 04:30:33.695109 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="381071ee-0780-4ba9-a924-4da9ea4113b2" path="/var/lib/kubelet/pods/381071ee-0780-4ba9-a924-4da9ea4113b2/volumes" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.551854 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.696383 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-config-data\") pod \"a3706ecc-dda1-498f-8319-b10379ceccb9\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.696603 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm24r\" (UniqueName: \"kubernetes.io/projected/a3706ecc-dda1-498f-8319-b10379ceccb9-kube-api-access-tm24r\") pod \"a3706ecc-dda1-498f-8319-b10379ceccb9\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.696659 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-scripts\") pod \"a3706ecc-dda1-498f-8319-b10379ceccb9\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.696712 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-fernet-keys\") pod \"a3706ecc-dda1-498f-8319-b10379ceccb9\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.696735 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-credential-keys\") pod \"a3706ecc-dda1-498f-8319-b10379ceccb9\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.696771 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-combined-ca-bundle\") pod \"a3706ecc-dda1-498f-8319-b10379ceccb9\" (UID: \"a3706ecc-dda1-498f-8319-b10379ceccb9\") " Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.702622 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-scripts" (OuterVolumeSpecName: "scripts") pod "a3706ecc-dda1-498f-8319-b10379ceccb9" (UID: "a3706ecc-dda1-498f-8319-b10379ceccb9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.703435 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a3706ecc-dda1-498f-8319-b10379ceccb9" (UID: "a3706ecc-dda1-498f-8319-b10379ceccb9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.708393 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a3706ecc-dda1-498f-8319-b10379ceccb9" (UID: "a3706ecc-dda1-498f-8319-b10379ceccb9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.713165 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3706ecc-dda1-498f-8319-b10379ceccb9-kube-api-access-tm24r" (OuterVolumeSpecName: "kube-api-access-tm24r") pod "a3706ecc-dda1-498f-8319-b10379ceccb9" (UID: "a3706ecc-dda1-498f-8319-b10379ceccb9"). InnerVolumeSpecName "kube-api-access-tm24r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.727949 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-config-data" (OuterVolumeSpecName: "config-data") pod "a3706ecc-dda1-498f-8319-b10379ceccb9" (UID: "a3706ecc-dda1-498f-8319-b10379ceccb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.736156 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3706ecc-dda1-498f-8319-b10379ceccb9" (UID: "a3706ecc-dda1-498f-8319-b10379ceccb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.798486 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.798528 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.798540 4770 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.798552 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.798563 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3706ecc-dda1-498f-8319-b10379ceccb9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:34 crc kubenswrapper[4770]: I1004 04:30:34.798576 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm24r\" (UniqueName: \"kubernetes.io/projected/a3706ecc-dda1-498f-8319-b10379ceccb9-kube-api-access-tm24r\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.190211 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fkn6b" event={"ID":"a3706ecc-dda1-498f-8319-b10379ceccb9","Type":"ContainerDied","Data":"cfcb5dd7af4a356e567d04c0315d5489de9586ba3e24bc05394d11b66d698ed8"} Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.190608 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfcb5dd7af4a356e567d04c0315d5489de9586ba3e24bc05394d11b66d698ed8" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.190289 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fkn6b" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.287652 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5454b8cff9-nj62p"] Oct 04 04:30:35 crc kubenswrapper[4770]: E1004 04:30:35.288024 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381071ee-0780-4ba9-a924-4da9ea4113b2" containerName="init" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.288047 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="381071ee-0780-4ba9-a924-4da9ea4113b2" containerName="init" Oct 04 04:30:35 crc kubenswrapper[4770]: E1004 04:30:35.288075 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3706ecc-dda1-498f-8319-b10379ceccb9" containerName="keystone-bootstrap" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.288084 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3706ecc-dda1-498f-8319-b10379ceccb9" containerName="keystone-bootstrap" Oct 04 04:30:35 crc kubenswrapper[4770]: E1004 04:30:35.288101 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381071ee-0780-4ba9-a924-4da9ea4113b2" containerName="dnsmasq-dns" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.288110 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="381071ee-0780-4ba9-a924-4da9ea4113b2" containerName="dnsmasq-dns" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.288279 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="381071ee-0780-4ba9-a924-4da9ea4113b2" containerName="dnsmasq-dns" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.288299 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3706ecc-dda1-498f-8319-b10379ceccb9" containerName="keystone-bootstrap" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.288927 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.294242 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-whxb7" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.294643 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.294945 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.296796 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.309964 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5454b8cff9-nj62p"] Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.410688 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blck4\" (UniqueName: \"kubernetes.io/projected/15d1751b-3112-4f88-ab65-8bcf99e6e336-kube-api-access-blck4\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.410767 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-fernet-keys\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.410918 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-credential-keys\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.410969 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-combined-ca-bundle\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.411068 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-config-data\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.411123 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-scripts\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.513122 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blck4\" (UniqueName: \"kubernetes.io/projected/15d1751b-3112-4f88-ab65-8bcf99e6e336-kube-api-access-blck4\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.513183 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-fernet-keys\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.513239 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-credential-keys\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.513269 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-combined-ca-bundle\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.513299 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-config-data\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.513333 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-scripts\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.518379 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-combined-ca-bundle\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.518449 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-credential-keys\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.518561 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-fernet-keys\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.518759 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-scripts\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.519867 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15d1751b-3112-4f88-ab65-8bcf99e6e336-config-data\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.530177 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blck4\" (UniqueName: \"kubernetes.io/projected/15d1751b-3112-4f88-ab65-8bcf99e6e336-kube-api-access-blck4\") pod \"keystone-5454b8cff9-nj62p\" (UID: \"15d1751b-3112-4f88-ab65-8bcf99e6e336\") " pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:35 crc kubenswrapper[4770]: I1004 04:30:35.610472 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:36 crc kubenswrapper[4770]: I1004 04:30:36.054810 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5454b8cff9-nj62p"] Oct 04 04:30:36 crc kubenswrapper[4770]: I1004 04:30:36.205132 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5454b8cff9-nj62p" event={"ID":"15d1751b-3112-4f88-ab65-8bcf99e6e336","Type":"ContainerStarted","Data":"1824ec84b6f15a5a647cfeb3848b71b0478f3bacb9eb7cac9ff0054d8a8bf1a9"} Oct 04 04:30:37 crc kubenswrapper[4770]: I1004 04:30:37.219567 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5454b8cff9-nj62p" event={"ID":"15d1751b-3112-4f88-ab65-8bcf99e6e336","Type":"ContainerStarted","Data":"9c4e7fcf3410d5e230c7f66cbdffbd558f1360e65bfc1461f6a9d0db71ec5cee"} Oct 04 04:30:37 crc kubenswrapper[4770]: I1004 04:30:37.219962 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:30:37 crc kubenswrapper[4770]: I1004 04:30:37.246569 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5454b8cff9-nj62p" podStartSLOduration=2.246553898 podStartE2EDuration="2.246553898s" podCreationTimestamp="2025-10-04 04:30:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:30:37.244534085 +0000 UTC m=+5248.536543807" watchObservedRunningTime="2025-10-04 04:30:37.246553898 +0000 UTC m=+5248.538563610" Oct 04 04:31:01 crc kubenswrapper[4770]: I1004 04:31:01.795763 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:31:01 crc kubenswrapper[4770]: I1004 04:31:01.796414 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:31:07 crc kubenswrapper[4770]: I1004 04:31:07.062790 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5454b8cff9-nj62p" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.439971 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.441650 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.446072 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-ctzlz" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.446635 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.448507 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.467249 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config\") pod \"openstackclient\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.467389 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config-secret\") pod \"openstackclient\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.467484 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndmdr\" (UniqueName: \"kubernetes.io/projected/ee66ba43-4a84-42a4-80b3-db7ebba91411-kube-api-access-ndmdr\") pod \"openstackclient\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.507940 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.513201 4770 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee66ba43-4a84-42a4-80b3-db7ebba91411\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:31:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:31:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:31:10Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:31:10Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:5670f9e696b19b76695bea5d4f9c46ac6494d96282f094de1243d8d7a06453b2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndmdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:31:10Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": Pod \"openstackclient\" is invalid: metadata.uid: Invalid value: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\": field is immutable" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.515075 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.522147 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.533228 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 04:31:10 crc kubenswrapper[4770]: E1004 04:31:10.533269 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-ndmdr openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="ee66ba43-4a84-42a4-80b3-db7ebba91411" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.535653 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.537482 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.556506 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.570742 4770 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ee66ba43-4a84-42a4-80b3-db7ebba91411" podUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.572151 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndmdr\" (UniqueName: \"kubernetes.io/projected/ee66ba43-4a84-42a4-80b3-db7ebba91411-kube-api-access-ndmdr\") pod \"openstackclient\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.572231 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config\") pod \"openstackclient\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.572284 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqrxc\" (UniqueName: \"kubernetes.io/projected/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-kube-api-access-kqrxc\") pod \"openstackclient\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.572332 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config-secret\") pod \"openstackclient\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.572401 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config\") pod \"openstackclient\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.572473 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config-secret\") pod \"openstackclient\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.573572 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config\") pod \"openstackclient\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: E1004 04:31:10.575963 4770 projected.go:194] Error preparing data for projected volume kube-api-access-ndmdr for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (ee66ba43-4a84-42a4-80b3-db7ebba91411) does not match the UID in record. The object might have been deleted and then recreated Oct 04 04:31:10 crc kubenswrapper[4770]: E1004 04:31:10.576131 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ee66ba43-4a84-42a4-80b3-db7ebba91411-kube-api-access-ndmdr podName:ee66ba43-4a84-42a4-80b3-db7ebba91411 nodeName:}" failed. No retries permitted until 2025-10-04 04:31:11.076046731 +0000 UTC m=+5282.368056463 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-ndmdr" (UniqueName: "kubernetes.io/projected/ee66ba43-4a84-42a4-80b3-db7ebba91411-kube-api-access-ndmdr") pod "openstackclient" (UID: "ee66ba43-4a84-42a4-80b3-db7ebba91411") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (ee66ba43-4a84-42a4-80b3-db7ebba91411) does not match the UID in record. The object might have been deleted and then recreated Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.576773 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.580178 4770 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ee66ba43-4a84-42a4-80b3-db7ebba91411" podUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.584453 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config-secret\") pod \"openstackclient\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.682980 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config-secret\") pod \"ee66ba43-4a84-42a4-80b3-db7ebba91411\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.683195 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config\") pod \"ee66ba43-4a84-42a4-80b3-db7ebba91411\" (UID: \"ee66ba43-4a84-42a4-80b3-db7ebba91411\") " Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.683777 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config-secret\") pod \"openstackclient\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.683944 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config\") pod \"openstackclient\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.683999 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqrxc\" (UniqueName: \"kubernetes.io/projected/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-kube-api-access-kqrxc\") pod \"openstackclient\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.684090 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndmdr\" (UniqueName: \"kubernetes.io/projected/ee66ba43-4a84-42a4-80b3-db7ebba91411-kube-api-access-ndmdr\") on node \"crc\" DevicePath \"\"" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.689070 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config-secret\") pod \"openstackclient\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.689183 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ee66ba43-4a84-42a4-80b3-db7ebba91411" (UID: "ee66ba43-4a84-42a4-80b3-db7ebba91411"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.689793 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config\") pod \"openstackclient\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.692355 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ee66ba43-4a84-42a4-80b3-db7ebba91411" (UID: "ee66ba43-4a84-42a4-80b3-db7ebba91411"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.719074 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqrxc\" (UniqueName: \"kubernetes.io/projected/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-kube-api-access-kqrxc\") pod \"openstackclient\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " pod="openstack/openstackclient" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.785140 4770 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.785170 4770 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee66ba43-4a84-42a4-80b3-db7ebba91411-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:31:10 crc kubenswrapper[4770]: I1004 04:31:10.860747 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:31:11 crc kubenswrapper[4770]: I1004 04:31:11.300096 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:31:11 crc kubenswrapper[4770]: W1004 04:31:11.303635 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a43bcdb_c35a_4dac_9cbe_3188ceca6cc7.slice/crio-948d1df636246c5083ef0229a63dbcc584591d776210587bfbe67a4dcae78d9e WatchSource:0}: Error finding container 948d1df636246c5083ef0229a63dbcc584591d776210587bfbe67a4dcae78d9e: Status 404 returned error can't find the container with id 948d1df636246c5083ef0229a63dbcc584591d776210587bfbe67a4dcae78d9e Oct 04 04:31:11 crc kubenswrapper[4770]: I1004 04:31:11.568260 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7","Type":"ContainerStarted","Data":"bf7f4d11301a1c79acddbb5a17a300c5a528dc95b07e0d485e84fcadfde0a0b9"} Oct 04 04:31:11 crc kubenswrapper[4770]: I1004 04:31:11.568289 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:31:11 crc kubenswrapper[4770]: I1004 04:31:11.568315 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7","Type":"ContainerStarted","Data":"948d1df636246c5083ef0229a63dbcc584591d776210587bfbe67a4dcae78d9e"} Oct 04 04:31:11 crc kubenswrapper[4770]: I1004 04:31:11.592061 4770 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ee66ba43-4a84-42a4-80b3-db7ebba91411" podUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" Oct 04 04:31:11 crc kubenswrapper[4770]: I1004 04:31:11.595702 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.5956840799999998 podStartE2EDuration="1.59568408s" podCreationTimestamp="2025-10-04 04:31:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:31:11.589224391 +0000 UTC m=+5282.881234103" watchObservedRunningTime="2025-10-04 04:31:11.59568408 +0000 UTC m=+5282.887693812" Oct 04 04:31:11 crc kubenswrapper[4770]: I1004 04:31:11.648542 4770 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ee66ba43-4a84-42a4-80b3-db7ebba91411" podUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" Oct 04 04:31:11 crc kubenswrapper[4770]: I1004 04:31:11.686937 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee66ba43-4a84-42a4-80b3-db7ebba91411" path="/var/lib/kubelet/pods/ee66ba43-4a84-42a4-80b3-db7ebba91411/volumes" Oct 04 04:31:31 crc kubenswrapper[4770]: I1004 04:31:31.796067 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:31:31 crc kubenswrapper[4770]: I1004 04:31:31.796864 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:31:31 crc kubenswrapper[4770]: I1004 04:31:31.796938 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:31:31 crc kubenswrapper[4770]: I1004 04:31:31.797944 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:31:31 crc kubenswrapper[4770]: I1004 04:31:31.798059 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" gracePeriod=600 Oct 04 04:31:31 crc kubenswrapper[4770]: E1004 04:31:31.927761 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:31:32 crc kubenswrapper[4770]: I1004 04:31:32.778253 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" exitCode=0 Oct 04 04:31:32 crc kubenswrapper[4770]: I1004 04:31:32.778316 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5"} Oct 04 04:31:32 crc kubenswrapper[4770]: I1004 04:31:32.778657 4770 scope.go:117] "RemoveContainer" containerID="6ee9dd0aa38f827331a89fde2cd36be108fb6bc2a67c9f1e927459a0a521b3c6" Oct 04 04:31:32 crc kubenswrapper[4770]: I1004 04:31:32.779133 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:31:32 crc kubenswrapper[4770]: E1004 04:31:32.779405 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:31:48 crc kubenswrapper[4770]: I1004 04:31:48.674473 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:31:48 crc kubenswrapper[4770]: E1004 04:31:48.675561 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:32:01 crc kubenswrapper[4770]: I1004 04:32:01.673907 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:32:01 crc kubenswrapper[4770]: E1004 04:32:01.674713 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:32:14 crc kubenswrapper[4770]: I1004 04:32:14.673468 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:32:14 crc kubenswrapper[4770]: E1004 04:32:14.674237 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.280756 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-657t2"] Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.296966 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.327732 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-657t2"] Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.437784 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxfnf\" (UniqueName: \"kubernetes.io/projected/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-kube-api-access-dxfnf\") pod \"redhat-operators-657t2\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.437843 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-utilities\") pod \"redhat-operators-657t2\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.437895 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-catalog-content\") pod \"redhat-operators-657t2\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.538841 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxfnf\" (UniqueName: \"kubernetes.io/projected/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-kube-api-access-dxfnf\") pod \"redhat-operators-657t2\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.538970 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-utilities\") pod \"redhat-operators-657t2\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.539034 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-catalog-content\") pod \"redhat-operators-657t2\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.539450 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-utilities\") pod \"redhat-operators-657t2\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.539533 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-catalog-content\") pod \"redhat-operators-657t2\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.565759 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxfnf\" (UniqueName: \"kubernetes.io/projected/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-kube-api-access-dxfnf\") pod \"redhat-operators-657t2\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:24 crc kubenswrapper[4770]: I1004 04:32:24.635815 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:25 crc kubenswrapper[4770]: I1004 04:32:25.092307 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-657t2"] Oct 04 04:32:25 crc kubenswrapper[4770]: I1004 04:32:25.315786 4770 generic.go:334] "Generic (PLEG): container finished" podID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerID="f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55" exitCode=0 Oct 04 04:32:25 crc kubenswrapper[4770]: I1004 04:32:25.315838 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-657t2" event={"ID":"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b","Type":"ContainerDied","Data":"f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55"} Oct 04 04:32:25 crc kubenswrapper[4770]: I1004 04:32:25.316066 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-657t2" event={"ID":"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b","Type":"ContainerStarted","Data":"b73525e8b6e57d067808d95795b9d33e6e41dae8e6b96c5504b2e89cbda017b8"} Oct 04 04:32:26 crc kubenswrapper[4770]: I1004 04:32:26.326377 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-657t2" event={"ID":"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b","Type":"ContainerStarted","Data":"69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9"} Oct 04 04:32:27 crc kubenswrapper[4770]: I1004 04:32:27.335500 4770 generic.go:334] "Generic (PLEG): container finished" podID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerID="69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9" exitCode=0 Oct 04 04:32:27 crc kubenswrapper[4770]: I1004 04:32:27.335543 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-657t2" event={"ID":"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b","Type":"ContainerDied","Data":"69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9"} Oct 04 04:32:29 crc kubenswrapper[4770]: I1004 04:32:29.353808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-657t2" event={"ID":"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b","Type":"ContainerStarted","Data":"25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe"} Oct 04 04:32:29 crc kubenswrapper[4770]: I1004 04:32:29.382366 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-657t2" podStartSLOduration=2.510234398 podStartE2EDuration="5.382336444s" podCreationTimestamp="2025-10-04 04:32:24 +0000 UTC" firstStartedPulling="2025-10-04 04:32:25.3176832 +0000 UTC m=+5356.609692912" lastFinishedPulling="2025-10-04 04:32:28.189785226 +0000 UTC m=+5359.481794958" observedRunningTime="2025-10-04 04:32:29.374846868 +0000 UTC m=+5360.666856620" watchObservedRunningTime="2025-10-04 04:32:29.382336444 +0000 UTC m=+5360.674346166" Oct 04 04:32:29 crc kubenswrapper[4770]: I1004 04:32:29.677769 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:32:29 crc kubenswrapper[4770]: E1004 04:32:29.678226 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:32:34 crc kubenswrapper[4770]: I1004 04:32:34.636390 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:34 crc kubenswrapper[4770]: I1004 04:32:34.636811 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:34 crc kubenswrapper[4770]: I1004 04:32:34.707268 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:35 crc kubenswrapper[4770]: I1004 04:32:35.453188 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:35 crc kubenswrapper[4770]: I1004 04:32:35.504036 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-657t2"] Oct 04 04:32:37 crc kubenswrapper[4770]: I1004 04:32:37.419103 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-657t2" podUID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerName="registry-server" containerID="cri-o://25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe" gracePeriod=2 Oct 04 04:32:37 crc kubenswrapper[4770]: I1004 04:32:37.907702 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.096865 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxfnf\" (UniqueName: \"kubernetes.io/projected/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-kube-api-access-dxfnf\") pod \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.097097 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-utilities\") pod \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.097325 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-catalog-content\") pod \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\" (UID: \"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b\") " Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.098199 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-utilities" (OuterVolumeSpecName: "utilities") pod "3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" (UID: "3b5b07f9-faf4-47fa-b1c9-c5adebc0684b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.103531 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-kube-api-access-dxfnf" (OuterVolumeSpecName: "kube-api-access-dxfnf") pod "3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" (UID: "3b5b07f9-faf4-47fa-b1c9-c5adebc0684b"). InnerVolumeSpecName "kube-api-access-dxfnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.179621 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" (UID: "3b5b07f9-faf4-47fa-b1c9-c5adebc0684b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.199948 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.200037 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxfnf\" (UniqueName: \"kubernetes.io/projected/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-kube-api-access-dxfnf\") on node \"crc\" DevicePath \"\"" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.200066 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.429249 4770 generic.go:334] "Generic (PLEG): container finished" podID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerID="25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe" exitCode=0 Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.429579 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-657t2" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.429441 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-657t2" event={"ID":"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b","Type":"ContainerDied","Data":"25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe"} Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.429718 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-657t2" event={"ID":"3b5b07f9-faf4-47fa-b1c9-c5adebc0684b","Type":"ContainerDied","Data":"b73525e8b6e57d067808d95795b9d33e6e41dae8e6b96c5504b2e89cbda017b8"} Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.429745 4770 scope.go:117] "RemoveContainer" containerID="25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.463308 4770 scope.go:117] "RemoveContainer" containerID="69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.483848 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-657t2"] Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.490252 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-657t2"] Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.507375 4770 scope.go:117] "RemoveContainer" containerID="f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.540053 4770 scope.go:117] "RemoveContainer" containerID="25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe" Oct 04 04:32:38 crc kubenswrapper[4770]: E1004 04:32:38.540540 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe\": container with ID starting with 25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe not found: ID does not exist" containerID="25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.540600 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe"} err="failed to get container status \"25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe\": rpc error: code = NotFound desc = could not find container \"25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe\": container with ID starting with 25599e01070f4707b5541dbdbf0e13681498844c7bbcc7451181e3366d1239fe not found: ID does not exist" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.540629 4770 scope.go:117] "RemoveContainer" containerID="69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9" Oct 04 04:32:38 crc kubenswrapper[4770]: E1004 04:32:38.540953 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9\": container with ID starting with 69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9 not found: ID does not exist" containerID="69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.540993 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9"} err="failed to get container status \"69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9\": rpc error: code = NotFound desc = could not find container \"69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9\": container with ID starting with 69e9f6b3ad6fa7664d6243bc8c83d3621e52e6ad669155b492fb9ff85839ffb9 not found: ID does not exist" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.541041 4770 scope.go:117] "RemoveContainer" containerID="f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55" Oct 04 04:32:38 crc kubenswrapper[4770]: E1004 04:32:38.541388 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55\": container with ID starting with f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55 not found: ID does not exist" containerID="f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55" Oct 04 04:32:38 crc kubenswrapper[4770]: I1004 04:32:38.541428 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55"} err="failed to get container status \"f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55\": rpc error: code = NotFound desc = could not find container \"f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55\": container with ID starting with f8e8ec3047c860dfb9da6ed58647233158da48f7dacc99cb97944eb166f22d55 not found: ID does not exist" Oct 04 04:32:39 crc kubenswrapper[4770]: I1004 04:32:39.691467 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" path="/var/lib/kubelet/pods/3b5b07f9-faf4-47fa-b1c9-c5adebc0684b/volumes" Oct 04 04:32:40 crc kubenswrapper[4770]: I1004 04:32:40.674402 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:32:40 crc kubenswrapper[4770]: E1004 04:32:40.675033 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:32:52 crc kubenswrapper[4770]: I1004 04:32:52.673198 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:32:52 crc kubenswrapper[4770]: E1004 04:32:52.673913 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:32:58 crc kubenswrapper[4770]: I1004 04:32:58.832268 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-z58nf"] Oct 04 04:32:58 crc kubenswrapper[4770]: E1004 04:32:58.833240 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerName="extract-content" Oct 04 04:32:58 crc kubenswrapper[4770]: I1004 04:32:58.833257 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerName="extract-content" Oct 04 04:32:58 crc kubenswrapper[4770]: E1004 04:32:58.833276 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerName="extract-utilities" Oct 04 04:32:58 crc kubenswrapper[4770]: I1004 04:32:58.833285 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerName="extract-utilities" Oct 04 04:32:58 crc kubenswrapper[4770]: E1004 04:32:58.833310 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerName="registry-server" Oct 04 04:32:58 crc kubenswrapper[4770]: I1004 04:32:58.833321 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerName="registry-server" Oct 04 04:32:58 crc kubenswrapper[4770]: I1004 04:32:58.833543 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b5b07f9-faf4-47fa-b1c9-c5adebc0684b" containerName="registry-server" Oct 04 04:32:58 crc kubenswrapper[4770]: I1004 04:32:58.834245 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z58nf" Oct 04 04:32:58 crc kubenswrapper[4770]: I1004 04:32:58.851725 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-z58nf"] Oct 04 04:32:58 crc kubenswrapper[4770]: I1004 04:32:58.991800 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnpff\" (UniqueName: \"kubernetes.io/projected/bc0de4a6-c99b-46c8-95d5-713076a5ed08-kube-api-access-jnpff\") pod \"barbican-db-create-z58nf\" (UID: \"bc0de4a6-c99b-46c8-95d5-713076a5ed08\") " pod="openstack/barbican-db-create-z58nf" Oct 04 04:32:59 crc kubenswrapper[4770]: I1004 04:32:59.093869 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnpff\" (UniqueName: \"kubernetes.io/projected/bc0de4a6-c99b-46c8-95d5-713076a5ed08-kube-api-access-jnpff\") pod \"barbican-db-create-z58nf\" (UID: \"bc0de4a6-c99b-46c8-95d5-713076a5ed08\") " pod="openstack/barbican-db-create-z58nf" Oct 04 04:32:59 crc kubenswrapper[4770]: I1004 04:32:59.120412 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnpff\" (UniqueName: \"kubernetes.io/projected/bc0de4a6-c99b-46c8-95d5-713076a5ed08-kube-api-access-jnpff\") pod \"barbican-db-create-z58nf\" (UID: \"bc0de4a6-c99b-46c8-95d5-713076a5ed08\") " pod="openstack/barbican-db-create-z58nf" Oct 04 04:32:59 crc kubenswrapper[4770]: I1004 04:32:59.167492 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z58nf" Oct 04 04:32:59 crc kubenswrapper[4770]: I1004 04:32:59.621684 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-z58nf"] Oct 04 04:32:59 crc kubenswrapper[4770]: W1004 04:32:59.635781 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc0de4a6_c99b_46c8_95d5_713076a5ed08.slice/crio-cc49e3298b90b22c12bc182668c2ccdc3a2fe9019b300a357ec37d8e60bfbd58 WatchSource:0}: Error finding container cc49e3298b90b22c12bc182668c2ccdc3a2fe9019b300a357ec37d8e60bfbd58: Status 404 returned error can't find the container with id cc49e3298b90b22c12bc182668c2ccdc3a2fe9019b300a357ec37d8e60bfbd58 Oct 04 04:33:00 crc kubenswrapper[4770]: I1004 04:33:00.653101 4770 generic.go:334] "Generic (PLEG): container finished" podID="bc0de4a6-c99b-46c8-95d5-713076a5ed08" containerID="f11cb846bf2694cc3030034642a355c37a0e5ac1410cde3c1a4c2a66ccf0f726" exitCode=0 Oct 04 04:33:00 crc kubenswrapper[4770]: I1004 04:33:00.653192 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z58nf" event={"ID":"bc0de4a6-c99b-46c8-95d5-713076a5ed08","Type":"ContainerDied","Data":"f11cb846bf2694cc3030034642a355c37a0e5ac1410cde3c1a4c2a66ccf0f726"} Oct 04 04:33:00 crc kubenswrapper[4770]: I1004 04:33:00.653454 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z58nf" event={"ID":"bc0de4a6-c99b-46c8-95d5-713076a5ed08","Type":"ContainerStarted","Data":"cc49e3298b90b22c12bc182668c2ccdc3a2fe9019b300a357ec37d8e60bfbd58"} Oct 04 04:33:01 crc kubenswrapper[4770]: I1004 04:33:01.986725 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z58nf" Oct 04 04:33:02 crc kubenswrapper[4770]: I1004 04:33:02.067125 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnpff\" (UniqueName: \"kubernetes.io/projected/bc0de4a6-c99b-46c8-95d5-713076a5ed08-kube-api-access-jnpff\") pod \"bc0de4a6-c99b-46c8-95d5-713076a5ed08\" (UID: \"bc0de4a6-c99b-46c8-95d5-713076a5ed08\") " Oct 04 04:33:02 crc kubenswrapper[4770]: I1004 04:33:02.076364 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0de4a6-c99b-46c8-95d5-713076a5ed08-kube-api-access-jnpff" (OuterVolumeSpecName: "kube-api-access-jnpff") pod "bc0de4a6-c99b-46c8-95d5-713076a5ed08" (UID: "bc0de4a6-c99b-46c8-95d5-713076a5ed08"). InnerVolumeSpecName "kube-api-access-jnpff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:02 crc kubenswrapper[4770]: I1004 04:33:02.169878 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnpff\" (UniqueName: \"kubernetes.io/projected/bc0de4a6-c99b-46c8-95d5-713076a5ed08-kube-api-access-jnpff\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:02 crc kubenswrapper[4770]: I1004 04:33:02.674151 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z58nf" event={"ID":"bc0de4a6-c99b-46c8-95d5-713076a5ed08","Type":"ContainerDied","Data":"cc49e3298b90b22c12bc182668c2ccdc3a2fe9019b300a357ec37d8e60bfbd58"} Oct 04 04:33:02 crc kubenswrapper[4770]: I1004 04:33:02.674203 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc49e3298b90b22c12bc182668c2ccdc3a2fe9019b300a357ec37d8e60bfbd58" Oct 04 04:33:02 crc kubenswrapper[4770]: I1004 04:33:02.674224 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z58nf" Oct 04 04:33:07 crc kubenswrapper[4770]: I1004 04:33:07.673965 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:33:07 crc kubenswrapper[4770]: E1004 04:33:07.674762 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:33:08 crc kubenswrapper[4770]: I1004 04:33:08.972642 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-eef1-account-create-6cp4q"] Oct 04 04:33:08 crc kubenswrapper[4770]: E1004 04:33:08.973095 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0de4a6-c99b-46c8-95d5-713076a5ed08" containerName="mariadb-database-create" Oct 04 04:33:08 crc kubenswrapper[4770]: I1004 04:33:08.973116 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0de4a6-c99b-46c8-95d5-713076a5ed08" containerName="mariadb-database-create" Oct 04 04:33:08 crc kubenswrapper[4770]: I1004 04:33:08.973379 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc0de4a6-c99b-46c8-95d5-713076a5ed08" containerName="mariadb-database-create" Oct 04 04:33:08 crc kubenswrapper[4770]: I1004 04:33:08.974102 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eef1-account-create-6cp4q" Oct 04 04:33:08 crc kubenswrapper[4770]: I1004 04:33:08.981489 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-eef1-account-create-6cp4q"] Oct 04 04:33:08 crc kubenswrapper[4770]: I1004 04:33:08.983934 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 04:33:09 crc kubenswrapper[4770]: I1004 04:33:09.095290 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9dgp\" (UniqueName: \"kubernetes.io/projected/42a97b61-8706-46df-847c-5c5a5e3528c2-kube-api-access-s9dgp\") pod \"barbican-eef1-account-create-6cp4q\" (UID: \"42a97b61-8706-46df-847c-5c5a5e3528c2\") " pod="openstack/barbican-eef1-account-create-6cp4q" Oct 04 04:33:09 crc kubenswrapper[4770]: I1004 04:33:09.196870 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9dgp\" (UniqueName: \"kubernetes.io/projected/42a97b61-8706-46df-847c-5c5a5e3528c2-kube-api-access-s9dgp\") pod \"barbican-eef1-account-create-6cp4q\" (UID: \"42a97b61-8706-46df-847c-5c5a5e3528c2\") " pod="openstack/barbican-eef1-account-create-6cp4q" Oct 04 04:33:09 crc kubenswrapper[4770]: I1004 04:33:09.217479 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9dgp\" (UniqueName: \"kubernetes.io/projected/42a97b61-8706-46df-847c-5c5a5e3528c2-kube-api-access-s9dgp\") pod \"barbican-eef1-account-create-6cp4q\" (UID: \"42a97b61-8706-46df-847c-5c5a5e3528c2\") " pod="openstack/barbican-eef1-account-create-6cp4q" Oct 04 04:33:09 crc kubenswrapper[4770]: I1004 04:33:09.298540 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eef1-account-create-6cp4q" Oct 04 04:33:09 crc kubenswrapper[4770]: I1004 04:33:09.759120 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-eef1-account-create-6cp4q"] Oct 04 04:33:09 crc kubenswrapper[4770]: I1004 04:33:09.764334 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 04:33:10 crc kubenswrapper[4770]: I1004 04:33:10.752933 4770 generic.go:334] "Generic (PLEG): container finished" podID="42a97b61-8706-46df-847c-5c5a5e3528c2" containerID="77f09aa65fb939fc11c487a147586f640f73308844864de1e7d50f18f101bc74" exitCode=0 Oct 04 04:33:10 crc kubenswrapper[4770]: I1004 04:33:10.753117 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-eef1-account-create-6cp4q" event={"ID":"42a97b61-8706-46df-847c-5c5a5e3528c2","Type":"ContainerDied","Data":"77f09aa65fb939fc11c487a147586f640f73308844864de1e7d50f18f101bc74"} Oct 04 04:33:10 crc kubenswrapper[4770]: I1004 04:33:10.753201 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-eef1-account-create-6cp4q" event={"ID":"42a97b61-8706-46df-847c-5c5a5e3528c2","Type":"ContainerStarted","Data":"79338876f5c2c12dd29fdda038d1f176db8c73c80169ab2f87a7a0ed88e26d10"} Oct 04 04:33:12 crc kubenswrapper[4770]: I1004 04:33:12.120552 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eef1-account-create-6cp4q" Oct 04 04:33:12 crc kubenswrapper[4770]: I1004 04:33:12.267295 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9dgp\" (UniqueName: \"kubernetes.io/projected/42a97b61-8706-46df-847c-5c5a5e3528c2-kube-api-access-s9dgp\") pod \"42a97b61-8706-46df-847c-5c5a5e3528c2\" (UID: \"42a97b61-8706-46df-847c-5c5a5e3528c2\") " Oct 04 04:33:12 crc kubenswrapper[4770]: I1004 04:33:12.273809 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a97b61-8706-46df-847c-5c5a5e3528c2-kube-api-access-s9dgp" (OuterVolumeSpecName: "kube-api-access-s9dgp") pod "42a97b61-8706-46df-847c-5c5a5e3528c2" (UID: "42a97b61-8706-46df-847c-5c5a5e3528c2"). InnerVolumeSpecName "kube-api-access-s9dgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:12 crc kubenswrapper[4770]: I1004 04:33:12.370153 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9dgp\" (UniqueName: \"kubernetes.io/projected/42a97b61-8706-46df-847c-5c5a5e3528c2-kube-api-access-s9dgp\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:12 crc kubenswrapper[4770]: I1004 04:33:12.777638 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-eef1-account-create-6cp4q" event={"ID":"42a97b61-8706-46df-847c-5c5a5e3528c2","Type":"ContainerDied","Data":"79338876f5c2c12dd29fdda038d1f176db8c73c80169ab2f87a7a0ed88e26d10"} Oct 04 04:33:12 crc kubenswrapper[4770]: I1004 04:33:12.777716 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79338876f5c2c12dd29fdda038d1f176db8c73c80169ab2f87a7a0ed88e26d10" Oct 04 04:33:12 crc kubenswrapper[4770]: I1004 04:33:12.777730 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eef1-account-create-6cp4q" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.241216 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-jxs7r"] Oct 04 04:33:14 crc kubenswrapper[4770]: E1004 04:33:14.241868 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a97b61-8706-46df-847c-5c5a5e3528c2" containerName="mariadb-account-create" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.241883 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a97b61-8706-46df-847c-5c5a5e3528c2" containerName="mariadb-account-create" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.242066 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="42a97b61-8706-46df-847c-5c5a5e3528c2" containerName="mariadb-account-create" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.242615 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.244634 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9zdw9" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.246432 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.255598 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jxs7r"] Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.403026 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-combined-ca-bundle\") pod \"barbican-db-sync-jxs7r\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.403142 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv8bw\" (UniqueName: \"kubernetes.io/projected/6a1a1ae0-84f3-4b49-84e5-1f1737410357-kube-api-access-cv8bw\") pod \"barbican-db-sync-jxs7r\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.403202 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-db-sync-config-data\") pod \"barbican-db-sync-jxs7r\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.504130 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-combined-ca-bundle\") pod \"barbican-db-sync-jxs7r\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.504257 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv8bw\" (UniqueName: \"kubernetes.io/projected/6a1a1ae0-84f3-4b49-84e5-1f1737410357-kube-api-access-cv8bw\") pod \"barbican-db-sync-jxs7r\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.504335 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-db-sync-config-data\") pod \"barbican-db-sync-jxs7r\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.509095 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-combined-ca-bundle\") pod \"barbican-db-sync-jxs7r\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.510867 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-db-sync-config-data\") pod \"barbican-db-sync-jxs7r\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.523843 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv8bw\" (UniqueName: \"kubernetes.io/projected/6a1a1ae0-84f3-4b49-84e5-1f1737410357-kube-api-access-cv8bw\") pod \"barbican-db-sync-jxs7r\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.573377 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:14 crc kubenswrapper[4770]: I1004 04:33:14.997446 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jxs7r"] Oct 04 04:33:15 crc kubenswrapper[4770]: I1004 04:33:15.818695 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jxs7r" event={"ID":"6a1a1ae0-84f3-4b49-84e5-1f1737410357","Type":"ContainerStarted","Data":"8eacfd4e7bd934d5b1e73d8573720bc8711fe1243b893829406c291b13f18666"} Oct 04 04:33:15 crc kubenswrapper[4770]: I1004 04:33:15.820223 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jxs7r" event={"ID":"6a1a1ae0-84f3-4b49-84e5-1f1737410357","Type":"ContainerStarted","Data":"bcfda2345260e0ff199fe1e66a552050d13a8e04698f09eb51d466d93deeadec"} Oct 04 04:33:15 crc kubenswrapper[4770]: I1004 04:33:15.838479 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-jxs7r" podStartSLOduration=1.838461343 podStartE2EDuration="1.838461343s" podCreationTimestamp="2025-10-04 04:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:33:15.833335258 +0000 UTC m=+5407.125344970" watchObservedRunningTime="2025-10-04 04:33:15.838461343 +0000 UTC m=+5407.130471055" Oct 04 04:33:16 crc kubenswrapper[4770]: I1004 04:33:16.830862 4770 generic.go:334] "Generic (PLEG): container finished" podID="6a1a1ae0-84f3-4b49-84e5-1f1737410357" containerID="8eacfd4e7bd934d5b1e73d8573720bc8711fe1243b893829406c291b13f18666" exitCode=0 Oct 04 04:33:16 crc kubenswrapper[4770]: I1004 04:33:16.830961 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jxs7r" event={"ID":"6a1a1ae0-84f3-4b49-84e5-1f1737410357","Type":"ContainerDied","Data":"8eacfd4e7bd934d5b1e73d8573720bc8711fe1243b893829406c291b13f18666"} Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.206162 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.370323 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv8bw\" (UniqueName: \"kubernetes.io/projected/6a1a1ae0-84f3-4b49-84e5-1f1737410357-kube-api-access-cv8bw\") pod \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.370404 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-db-sync-config-data\") pod \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.370456 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-combined-ca-bundle\") pod \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\" (UID: \"6a1a1ae0-84f3-4b49-84e5-1f1737410357\") " Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.376163 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6a1a1ae0-84f3-4b49-84e5-1f1737410357" (UID: "6a1a1ae0-84f3-4b49-84e5-1f1737410357"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.376297 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a1a1ae0-84f3-4b49-84e5-1f1737410357-kube-api-access-cv8bw" (OuterVolumeSpecName: "kube-api-access-cv8bw") pod "6a1a1ae0-84f3-4b49-84e5-1f1737410357" (UID: "6a1a1ae0-84f3-4b49-84e5-1f1737410357"). InnerVolumeSpecName "kube-api-access-cv8bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.396231 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a1a1ae0-84f3-4b49-84e5-1f1737410357" (UID: "6a1a1ae0-84f3-4b49-84e5-1f1737410357"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.473341 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv8bw\" (UniqueName: \"kubernetes.io/projected/6a1a1ae0-84f3-4b49-84e5-1f1737410357-kube-api-access-cv8bw\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.473422 4770 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.473436 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1a1ae0-84f3-4b49-84e5-1f1737410357-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.852323 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jxs7r" event={"ID":"6a1a1ae0-84f3-4b49-84e5-1f1737410357","Type":"ContainerDied","Data":"bcfda2345260e0ff199fe1e66a552050d13a8e04698f09eb51d466d93deeadec"} Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.852390 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcfda2345260e0ff199fe1e66a552050d13a8e04698f09eb51d466d93deeadec" Oct 04 04:33:18 crc kubenswrapper[4770]: I1004 04:33:18.852401 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jxs7r" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.156739 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5fc6d5ddf-dhcsd"] Oct 04 04:33:19 crc kubenswrapper[4770]: E1004 04:33:19.157409 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1a1ae0-84f3-4b49-84e5-1f1737410357" containerName="barbican-db-sync" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.157426 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1a1ae0-84f3-4b49-84e5-1f1737410357" containerName="barbican-db-sync" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.157602 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1a1ae0-84f3-4b49-84e5-1f1737410357" containerName="barbican-db-sync" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.158491 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.168878 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.169526 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.169671 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9zdw9" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.184064 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5fc6d5ddf-dhcsd"] Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.194548 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-665bfd56dd-lf8wl"] Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.207123 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.215390 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.244224 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-665bfd56dd-lf8wl"] Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.269065 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56b4bdb47c-fgzf7"] Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.270455 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.283455 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56b4bdb47c-fgzf7"] Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.287883 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-logs\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.287923 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-config-data-custom\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.287955 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4839fba-4fcd-4b16-a38a-af41ed293efc-config-data\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.287976 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4839fba-4fcd-4b16-a38a-af41ed293efc-logs\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.287999 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4839fba-4fcd-4b16-a38a-af41ed293efc-combined-ca-bundle\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.288037 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pszhb\" (UniqueName: \"kubernetes.io/projected/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-kube-api-access-pszhb\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.288054 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-config-data\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.288079 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-combined-ca-bundle\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.288147 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4839fba-4fcd-4b16-a38a-af41ed293efc-config-data-custom\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.288197 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcmt4\" (UniqueName: \"kubernetes.io/projected/d4839fba-4fcd-4b16-a38a-af41ed293efc-kube-api-access-hcmt4\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.339850 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-54dd7bd464-g9bnt"] Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.341555 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.343354 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.352641 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54dd7bd464-g9bnt"] Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.389837 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4839fba-4fcd-4b16-a38a-af41ed293efc-config-data-custom\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.389902 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-dns-svc\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.389975 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-sb\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390087 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-config\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390111 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fcjf\" (UniqueName: \"kubernetes.io/projected/969bbf54-1ba4-484f-84df-10521d8c2b4f-kube-api-access-7fcjf\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390138 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcmt4\" (UniqueName: \"kubernetes.io/projected/d4839fba-4fcd-4b16-a38a-af41ed293efc-kube-api-access-hcmt4\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390175 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-logs\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390194 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-config-data-custom\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390221 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4839fba-4fcd-4b16-a38a-af41ed293efc-config-data\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390237 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4839fba-4fcd-4b16-a38a-af41ed293efc-logs\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390256 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4839fba-4fcd-4b16-a38a-af41ed293efc-combined-ca-bundle\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390279 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pszhb\" (UniqueName: \"kubernetes.io/projected/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-kube-api-access-pszhb\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390295 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-config-data\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390314 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-combined-ca-bundle\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390345 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-nb\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.390997 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4839fba-4fcd-4b16-a38a-af41ed293efc-logs\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.391563 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-logs\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.408535 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-config-data\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.409886 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4839fba-4fcd-4b16-a38a-af41ed293efc-combined-ca-bundle\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.413550 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-combined-ca-bundle\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.417851 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4839fba-4fcd-4b16-a38a-af41ed293efc-config-data-custom\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.418413 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4839fba-4fcd-4b16-a38a-af41ed293efc-config-data\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.419162 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-config-data-custom\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.429706 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pszhb\" (UniqueName: \"kubernetes.io/projected/c73f3d35-51e8-4ad9-be44-78ae4cd1fec1-kube-api-access-pszhb\") pod \"barbican-worker-5fc6d5ddf-dhcsd\" (UID: \"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1\") " pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.429825 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcmt4\" (UniqueName: \"kubernetes.io/projected/d4839fba-4fcd-4b16-a38a-af41ed293efc-kube-api-access-hcmt4\") pod \"barbican-keystone-listener-665bfd56dd-lf8wl\" (UID: \"d4839fba-4fcd-4b16-a38a-af41ed293efc\") " pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492071 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-combined-ca-bundle\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492186 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-config-data-custom\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492217 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p75jr\" (UniqueName: \"kubernetes.io/projected/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-kube-api-access-p75jr\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492345 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-nb\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492442 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-logs\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492484 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-dns-svc\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492538 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-sb\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492580 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-config-data\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492615 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-config\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.492646 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fcjf\" (UniqueName: \"kubernetes.io/projected/969bbf54-1ba4-484f-84df-10521d8c2b4f-kube-api-access-7fcjf\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.493536 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-dns-svc\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.493925 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-sb\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.494146 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-nb\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.494614 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-config\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.507718 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.510941 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fcjf\" (UniqueName: \"kubernetes.io/projected/969bbf54-1ba4-484f-84df-10521d8c2b4f-kube-api-access-7fcjf\") pod \"dnsmasq-dns-56b4bdb47c-fgzf7\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.530497 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.594154 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-combined-ca-bundle\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.594228 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-config-data-custom\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.594251 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p75jr\" (UniqueName: \"kubernetes.io/projected/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-kube-api-access-p75jr\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.594288 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-logs\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.594327 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-config-data\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.596089 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-logs\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.598236 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-config-data\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.598294 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-config-data-custom\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.603190 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-combined-ca-bundle\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.603605 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.630501 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p75jr\" (UniqueName: \"kubernetes.io/projected/dedbb8f2-13d7-4969-9e03-f24f613f0f0a-kube-api-access-p75jr\") pod \"barbican-api-54dd7bd464-g9bnt\" (UID: \"dedbb8f2-13d7-4969-9e03-f24f613f0f0a\") " pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:19 crc kubenswrapper[4770]: I1004 04:33:19.658486 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:20 crc kubenswrapper[4770]: W1004 04:33:20.022423 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4839fba_4fcd_4b16_a38a_af41ed293efc.slice/crio-aebeb00cb2b340b90cbbf60e2a8eb9051b68f78d031043a2ad8f4d0e78a3cc46 WatchSource:0}: Error finding container aebeb00cb2b340b90cbbf60e2a8eb9051b68f78d031043a2ad8f4d0e78a3cc46: Status 404 returned error can't find the container with id aebeb00cb2b340b90cbbf60e2a8eb9051b68f78d031043a2ad8f4d0e78a3cc46 Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.022740 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-665bfd56dd-lf8wl"] Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.119592 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5fc6d5ddf-dhcsd"] Oct 04 04:33:20 crc kubenswrapper[4770]: W1004 04:33:20.125914 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc73f3d35_51e8_4ad9_be44_78ae4cd1fec1.slice/crio-2c4353e7dd3016368dd60113823ae03c1e013d9205c812546389d72ee47521c2 WatchSource:0}: Error finding container 2c4353e7dd3016368dd60113823ae03c1e013d9205c812546389d72ee47521c2: Status 404 returned error can't find the container with id 2c4353e7dd3016368dd60113823ae03c1e013d9205c812546389d72ee47521c2 Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.209894 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56b4bdb47c-fgzf7"] Oct 04 04:33:20 crc kubenswrapper[4770]: W1004 04:33:20.219399 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod969bbf54_1ba4_484f_84df_10521d8c2b4f.slice/crio-9bb617c3c2e2e4f916bd41e880be5a3781f317fcb5a1abd434e363793f40b1dd WatchSource:0}: Error finding container 9bb617c3c2e2e4f916bd41e880be5a3781f317fcb5a1abd434e363793f40b1dd: Status 404 returned error can't find the container with id 9bb617c3c2e2e4f916bd41e880be5a3781f317fcb5a1abd434e363793f40b1dd Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.223144 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54dd7bd464-g9bnt"] Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.876226 4770 generic.go:334] "Generic (PLEG): container finished" podID="969bbf54-1ba4-484f-84df-10521d8c2b4f" containerID="6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6" exitCode=0 Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.876369 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" event={"ID":"969bbf54-1ba4-484f-84df-10521d8c2b4f","Type":"ContainerDied","Data":"6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.876649 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" event={"ID":"969bbf54-1ba4-484f-84df-10521d8c2b4f","Type":"ContainerStarted","Data":"9bb617c3c2e2e4f916bd41e880be5a3781f317fcb5a1abd434e363793f40b1dd"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.879707 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" event={"ID":"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1","Type":"ContainerStarted","Data":"183b9075e48d1bfdda77d672d96b6909c0a1f29f46f96f7b1a3038e0f92474e6"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.879744 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" event={"ID":"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1","Type":"ContainerStarted","Data":"1bf6ca33a061a82989482ae79a73a38f71c299f40d5c6ba5dad9aaaaef025fd7"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.879756 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" event={"ID":"c73f3d35-51e8-4ad9-be44-78ae4cd1fec1","Type":"ContainerStarted","Data":"2c4353e7dd3016368dd60113823ae03c1e013d9205c812546389d72ee47521c2"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.896939 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54dd7bd464-g9bnt" event={"ID":"dedbb8f2-13d7-4969-9e03-f24f613f0f0a","Type":"ContainerStarted","Data":"40a2983cfbae0c8cfe387ed2173548bdceb5b969453bb988bc6e02860fd92e85"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.897269 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54dd7bd464-g9bnt" event={"ID":"dedbb8f2-13d7-4969-9e03-f24f613f0f0a","Type":"ContainerStarted","Data":"b1079a386adac6fb35034e183b83d78b93db4fd6023a94d4357669569e343feb"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.900808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" event={"ID":"d4839fba-4fcd-4b16-a38a-af41ed293efc","Type":"ContainerStarted","Data":"730e41760780231a93a74f635c17f5a334d4029e5e1124a74513e1bb71c1e97e"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.900869 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" event={"ID":"d4839fba-4fcd-4b16-a38a-af41ed293efc","Type":"ContainerStarted","Data":"8acd9b338124b0634b1f150a8c3b63ffe89ac61fae2884ab0e1f48dc60274286"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.900879 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" event={"ID":"d4839fba-4fcd-4b16-a38a-af41ed293efc","Type":"ContainerStarted","Data":"aebeb00cb2b340b90cbbf60e2a8eb9051b68f78d031043a2ad8f4d0e78a3cc46"} Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.938255 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5fc6d5ddf-dhcsd" podStartSLOduration=1.93820557 podStartE2EDuration="1.93820557s" podCreationTimestamp="2025-10-04 04:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:33:20.913790162 +0000 UTC m=+5412.205799884" watchObservedRunningTime="2025-10-04 04:33:20.93820557 +0000 UTC m=+5412.230215302" Oct 04 04:33:20 crc kubenswrapper[4770]: I1004 04:33:20.964686 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-665bfd56dd-lf8wl" podStartSLOduration=1.9646641009999999 podStartE2EDuration="1.964664101s" podCreationTimestamp="2025-10-04 04:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:33:20.932222134 +0000 UTC m=+5412.224231866" watchObservedRunningTime="2025-10-04 04:33:20.964664101 +0000 UTC m=+5412.256673823" Oct 04 04:33:21 crc kubenswrapper[4770]: I1004 04:33:21.674288 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:33:21 crc kubenswrapper[4770]: E1004 04:33:21.674861 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:33:21 crc kubenswrapper[4770]: I1004 04:33:21.909851 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54dd7bd464-g9bnt" event={"ID":"dedbb8f2-13d7-4969-9e03-f24f613f0f0a","Type":"ContainerStarted","Data":"8f35d78ce4196e2bf62c2ca823235cccae869e49965a6005fd05c49a85695193"} Oct 04 04:33:21 crc kubenswrapper[4770]: I1004 04:33:21.909950 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:21 crc kubenswrapper[4770]: I1004 04:33:21.910051 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:21 crc kubenswrapper[4770]: I1004 04:33:21.912582 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" event={"ID":"969bbf54-1ba4-484f-84df-10521d8c2b4f","Type":"ContainerStarted","Data":"898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab"} Oct 04 04:33:21 crc kubenswrapper[4770]: I1004 04:33:21.937265 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-54dd7bd464-g9bnt" podStartSLOduration=2.937238321 podStartE2EDuration="2.937238321s" podCreationTimestamp="2025-10-04 04:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:33:21.933097543 +0000 UTC m=+5413.225107265" watchObservedRunningTime="2025-10-04 04:33:21.937238321 +0000 UTC m=+5413.229248033" Oct 04 04:33:22 crc kubenswrapper[4770]: I1004 04:33:22.924908 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:29 crc kubenswrapper[4770]: I1004 04:33:29.605687 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:33:29 crc kubenswrapper[4770]: I1004 04:33:29.627332 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" podStartSLOduration=10.627315849 podStartE2EDuration="10.627315849s" podCreationTimestamp="2025-10-04 04:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:33:21.959919474 +0000 UTC m=+5413.251929216" watchObservedRunningTime="2025-10-04 04:33:29.627315849 +0000 UTC m=+5420.919325561" Oct 04 04:33:29 crc kubenswrapper[4770]: I1004 04:33:29.668580 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59f4ccb565-qspsj"] Oct 04 04:33:29 crc kubenswrapper[4770]: I1004 04:33:29.668839 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" podUID="cb075f34-9b51-467d-978c-36c563b48d81" containerName="dnsmasq-dns" containerID="cri-o://116bce903fb7391ca951e960857439b3fbff228a9dcc1a3956168bba9e9f8871" gracePeriod=10 Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.006820 4770 generic.go:334] "Generic (PLEG): container finished" podID="cb075f34-9b51-467d-978c-36c563b48d81" containerID="116bce903fb7391ca951e960857439b3fbff228a9dcc1a3956168bba9e9f8871" exitCode=0 Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.006998 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" event={"ID":"cb075f34-9b51-467d-978c-36c563b48d81","Type":"ContainerDied","Data":"116bce903fb7391ca951e960857439b3fbff228a9dcc1a3956168bba9e9f8871"} Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.206528 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.301897 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-nb\") pod \"cb075f34-9b51-467d-978c-36c563b48d81\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.301956 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-dns-svc\") pod \"cb075f34-9b51-467d-978c-36c563b48d81\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.301989 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-sb\") pod \"cb075f34-9b51-467d-978c-36c563b48d81\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.302276 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-config\") pod \"cb075f34-9b51-467d-978c-36c563b48d81\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.302366 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9jfc\" (UniqueName: \"kubernetes.io/projected/cb075f34-9b51-467d-978c-36c563b48d81-kube-api-access-f9jfc\") pod \"cb075f34-9b51-467d-978c-36c563b48d81\" (UID: \"cb075f34-9b51-467d-978c-36c563b48d81\") " Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.307603 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb075f34-9b51-467d-978c-36c563b48d81-kube-api-access-f9jfc" (OuterVolumeSpecName: "kube-api-access-f9jfc") pod "cb075f34-9b51-467d-978c-36c563b48d81" (UID: "cb075f34-9b51-467d-978c-36c563b48d81"). InnerVolumeSpecName "kube-api-access-f9jfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.342766 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cb075f34-9b51-467d-978c-36c563b48d81" (UID: "cb075f34-9b51-467d-978c-36c563b48d81"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.343787 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cb075f34-9b51-467d-978c-36c563b48d81" (UID: "cb075f34-9b51-467d-978c-36c563b48d81"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.346352 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cb075f34-9b51-467d-978c-36c563b48d81" (UID: "cb075f34-9b51-467d-978c-36c563b48d81"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.351094 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-config" (OuterVolumeSpecName: "config") pod "cb075f34-9b51-467d-978c-36c563b48d81" (UID: "cb075f34-9b51-467d-978c-36c563b48d81"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.404373 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.404414 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.404448 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.404459 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9jfc\" (UniqueName: \"kubernetes.io/projected/cb075f34-9b51-467d-978c-36c563b48d81-kube-api-access-f9jfc\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:30 crc kubenswrapper[4770]: I1004 04:33:30.404469 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb075f34-9b51-467d-978c-36c563b48d81-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:31 crc kubenswrapper[4770]: I1004 04:33:31.026591 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" event={"ID":"cb075f34-9b51-467d-978c-36c563b48d81","Type":"ContainerDied","Data":"3375981a47cee8018ce9760a8e3aeb469c681f534772ebc953e6b1f92641640f"} Oct 04 04:33:31 crc kubenswrapper[4770]: I1004 04:33:31.026654 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f4ccb565-qspsj" Oct 04 04:33:31 crc kubenswrapper[4770]: I1004 04:33:31.026947 4770 scope.go:117] "RemoveContainer" containerID="116bce903fb7391ca951e960857439b3fbff228a9dcc1a3956168bba9e9f8871" Oct 04 04:33:31 crc kubenswrapper[4770]: I1004 04:33:31.052396 4770 scope.go:117] "RemoveContainer" containerID="0f218d03b43bb2375c2357953820aa65de71bb15fa5eab443fdacd6a6f4a6e39" Oct 04 04:33:31 crc kubenswrapper[4770]: I1004 04:33:31.079703 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59f4ccb565-qspsj"] Oct 04 04:33:31 crc kubenswrapper[4770]: I1004 04:33:31.081071 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:31 crc kubenswrapper[4770]: I1004 04:33:31.088685 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59f4ccb565-qspsj"] Oct 04 04:33:31 crc kubenswrapper[4770]: I1004 04:33:31.126865 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54dd7bd464-g9bnt" Oct 04 04:33:31 crc kubenswrapper[4770]: I1004 04:33:31.688636 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb075f34-9b51-467d-978c-36c563b48d81" path="/var/lib/kubelet/pods/cb075f34-9b51-467d-978c-36c563b48d81/volumes" Oct 04 04:33:34 crc kubenswrapper[4770]: I1004 04:33:34.673515 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:33:34 crc kubenswrapper[4770]: E1004 04:33:34.674299 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.189347 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-d2nkz"] Oct 04 04:33:45 crc kubenswrapper[4770]: E1004 04:33:45.190536 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb075f34-9b51-467d-978c-36c563b48d81" containerName="init" Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.190551 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb075f34-9b51-467d-978c-36c563b48d81" containerName="init" Oct 04 04:33:45 crc kubenswrapper[4770]: E1004 04:33:45.190562 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb075f34-9b51-467d-978c-36c563b48d81" containerName="dnsmasq-dns" Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.190568 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb075f34-9b51-467d-978c-36c563b48d81" containerName="dnsmasq-dns" Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.190722 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb075f34-9b51-467d-978c-36c563b48d81" containerName="dnsmasq-dns" Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.191410 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-d2nkz" Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.196406 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-d2nkz"] Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.277125 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb89v\" (UniqueName: \"kubernetes.io/projected/70d9f4e9-231d-49fe-b7ac-ac9f9301e501-kube-api-access-bb89v\") pod \"neutron-db-create-d2nkz\" (UID: \"70d9f4e9-231d-49fe-b7ac-ac9f9301e501\") " pod="openstack/neutron-db-create-d2nkz" Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.378822 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb89v\" (UniqueName: \"kubernetes.io/projected/70d9f4e9-231d-49fe-b7ac-ac9f9301e501-kube-api-access-bb89v\") pod \"neutron-db-create-d2nkz\" (UID: \"70d9f4e9-231d-49fe-b7ac-ac9f9301e501\") " pod="openstack/neutron-db-create-d2nkz" Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.399482 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb89v\" (UniqueName: \"kubernetes.io/projected/70d9f4e9-231d-49fe-b7ac-ac9f9301e501-kube-api-access-bb89v\") pod \"neutron-db-create-d2nkz\" (UID: \"70d9f4e9-231d-49fe-b7ac-ac9f9301e501\") " pod="openstack/neutron-db-create-d2nkz" Oct 04 04:33:45 crc kubenswrapper[4770]: I1004 04:33:45.515563 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-d2nkz" Oct 04 04:33:46 crc kubenswrapper[4770]: I1004 04:33:46.010186 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-d2nkz"] Oct 04 04:33:46 crc kubenswrapper[4770]: I1004 04:33:46.164500 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-d2nkz" event={"ID":"70d9f4e9-231d-49fe-b7ac-ac9f9301e501","Type":"ContainerStarted","Data":"e4ab23cc4be0c5db40c532387ae5ff846f8642e28e3a238cef0c5df91b0ad748"} Oct 04 04:33:47 crc kubenswrapper[4770]: I1004 04:33:47.177817 4770 generic.go:334] "Generic (PLEG): container finished" podID="70d9f4e9-231d-49fe-b7ac-ac9f9301e501" containerID="76bc11893d6003cdc5a3567ab4d0e9505654d15686caad6c1128642d03b94d4f" exitCode=0 Oct 04 04:33:47 crc kubenswrapper[4770]: I1004 04:33:47.178147 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-d2nkz" event={"ID":"70d9f4e9-231d-49fe-b7ac-ac9f9301e501","Type":"ContainerDied","Data":"76bc11893d6003cdc5a3567ab4d0e9505654d15686caad6c1128642d03b94d4f"} Oct 04 04:33:48 crc kubenswrapper[4770]: I1004 04:33:48.524636 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-d2nkz" Oct 04 04:33:48 crc kubenswrapper[4770]: I1004 04:33:48.632508 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb89v\" (UniqueName: \"kubernetes.io/projected/70d9f4e9-231d-49fe-b7ac-ac9f9301e501-kube-api-access-bb89v\") pod \"70d9f4e9-231d-49fe-b7ac-ac9f9301e501\" (UID: \"70d9f4e9-231d-49fe-b7ac-ac9f9301e501\") " Oct 04 04:33:48 crc kubenswrapper[4770]: I1004 04:33:48.639553 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70d9f4e9-231d-49fe-b7ac-ac9f9301e501-kube-api-access-bb89v" (OuterVolumeSpecName: "kube-api-access-bb89v") pod "70d9f4e9-231d-49fe-b7ac-ac9f9301e501" (UID: "70d9f4e9-231d-49fe-b7ac-ac9f9301e501"). InnerVolumeSpecName "kube-api-access-bb89v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:48 crc kubenswrapper[4770]: I1004 04:33:48.687374 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:33:48 crc kubenswrapper[4770]: E1004 04:33:48.687812 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:33:48 crc kubenswrapper[4770]: I1004 04:33:48.734892 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb89v\" (UniqueName: \"kubernetes.io/projected/70d9f4e9-231d-49fe-b7ac-ac9f9301e501-kube-api-access-bb89v\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:49 crc kubenswrapper[4770]: I1004 04:33:49.197692 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-d2nkz" event={"ID":"70d9f4e9-231d-49fe-b7ac-ac9f9301e501","Type":"ContainerDied","Data":"e4ab23cc4be0c5db40c532387ae5ff846f8642e28e3a238cef0c5df91b0ad748"} Oct 04 04:33:49 crc kubenswrapper[4770]: I1004 04:33:49.198133 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4ab23cc4be0c5db40c532387ae5ff846f8642e28e3a238cef0c5df91b0ad748" Oct 04 04:33:49 crc kubenswrapper[4770]: I1004 04:33:49.197810 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-d2nkz" Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.266836 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f532-account-create-q7svr"] Oct 04 04:33:55 crc kubenswrapper[4770]: E1004 04:33:55.268491 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d9f4e9-231d-49fe-b7ac-ac9f9301e501" containerName="mariadb-database-create" Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.268515 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d9f4e9-231d-49fe-b7ac-ac9f9301e501" containerName="mariadb-database-create" Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.268801 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d9f4e9-231d-49fe-b7ac-ac9f9301e501" containerName="mariadb-database-create" Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.269691 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f532-account-create-q7svr" Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.273524 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.285420 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f532-account-create-q7svr"] Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.372710 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9jj6\" (UniqueName: \"kubernetes.io/projected/fc0a8eb2-d199-44ed-8dba-0b654c4261ca-kube-api-access-k9jj6\") pod \"neutron-f532-account-create-q7svr\" (UID: \"fc0a8eb2-d199-44ed-8dba-0b654c4261ca\") " pod="openstack/neutron-f532-account-create-q7svr" Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.473940 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9jj6\" (UniqueName: \"kubernetes.io/projected/fc0a8eb2-d199-44ed-8dba-0b654c4261ca-kube-api-access-k9jj6\") pod \"neutron-f532-account-create-q7svr\" (UID: \"fc0a8eb2-d199-44ed-8dba-0b654c4261ca\") " pod="openstack/neutron-f532-account-create-q7svr" Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.507455 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9jj6\" (UniqueName: \"kubernetes.io/projected/fc0a8eb2-d199-44ed-8dba-0b654c4261ca-kube-api-access-k9jj6\") pod \"neutron-f532-account-create-q7svr\" (UID: \"fc0a8eb2-d199-44ed-8dba-0b654c4261ca\") " pod="openstack/neutron-f532-account-create-q7svr" Oct 04 04:33:55 crc kubenswrapper[4770]: I1004 04:33:55.596237 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f532-account-create-q7svr" Oct 04 04:33:56 crc kubenswrapper[4770]: I1004 04:33:56.096528 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f532-account-create-q7svr"] Oct 04 04:33:56 crc kubenswrapper[4770]: I1004 04:33:56.260173 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f532-account-create-q7svr" event={"ID":"fc0a8eb2-d199-44ed-8dba-0b654c4261ca","Type":"ContainerStarted","Data":"6ae88219927e53344aa9b9c46b4470850d183a636d152a8c05ea8058ae3be798"} Oct 04 04:33:57 crc kubenswrapper[4770]: I1004 04:33:57.278609 4770 generic.go:334] "Generic (PLEG): container finished" podID="fc0a8eb2-d199-44ed-8dba-0b654c4261ca" containerID="ce09c6ff968307577b4a88a60a9bbc1c71a073fa3b1ab7881ef48a87abff3862" exitCode=0 Oct 04 04:33:57 crc kubenswrapper[4770]: I1004 04:33:57.278681 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f532-account-create-q7svr" event={"ID":"fc0a8eb2-d199-44ed-8dba-0b654c4261ca","Type":"ContainerDied","Data":"ce09c6ff968307577b4a88a60a9bbc1c71a073fa3b1ab7881ef48a87abff3862"} Oct 04 04:33:58 crc kubenswrapper[4770]: I1004 04:33:58.628599 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f532-account-create-q7svr" Oct 04 04:33:58 crc kubenswrapper[4770]: I1004 04:33:58.737360 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9jj6\" (UniqueName: \"kubernetes.io/projected/fc0a8eb2-d199-44ed-8dba-0b654c4261ca-kube-api-access-k9jj6\") pod \"fc0a8eb2-d199-44ed-8dba-0b654c4261ca\" (UID: \"fc0a8eb2-d199-44ed-8dba-0b654c4261ca\") " Oct 04 04:33:58 crc kubenswrapper[4770]: I1004 04:33:58.743866 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0a8eb2-d199-44ed-8dba-0b654c4261ca-kube-api-access-k9jj6" (OuterVolumeSpecName: "kube-api-access-k9jj6") pod "fc0a8eb2-d199-44ed-8dba-0b654c4261ca" (UID: "fc0a8eb2-d199-44ed-8dba-0b654c4261ca"). InnerVolumeSpecName "kube-api-access-k9jj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:58 crc kubenswrapper[4770]: I1004 04:33:58.839653 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9jj6\" (UniqueName: \"kubernetes.io/projected/fc0a8eb2-d199-44ed-8dba-0b654c4261ca-kube-api-access-k9jj6\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:59 crc kubenswrapper[4770]: I1004 04:33:59.304851 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f532-account-create-q7svr" event={"ID":"fc0a8eb2-d199-44ed-8dba-0b654c4261ca","Type":"ContainerDied","Data":"6ae88219927e53344aa9b9c46b4470850d183a636d152a8c05ea8058ae3be798"} Oct 04 04:33:59 crc kubenswrapper[4770]: I1004 04:33:59.304910 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ae88219927e53344aa9b9c46b4470850d183a636d152a8c05ea8058ae3be798" Oct 04 04:33:59 crc kubenswrapper[4770]: I1004 04:33:59.305240 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f532-account-create-q7svr" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.431587 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-dvh5k"] Oct 04 04:34:00 crc kubenswrapper[4770]: E1004 04:34:00.432370 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0a8eb2-d199-44ed-8dba-0b654c4261ca" containerName="mariadb-account-create" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.432387 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0a8eb2-d199-44ed-8dba-0b654c4261ca" containerName="mariadb-account-create" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.432583 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0a8eb2-d199-44ed-8dba-0b654c4261ca" containerName="mariadb-account-create" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.433279 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.436736 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5ll6j" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.436814 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.437158 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.439498 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dvh5k"] Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.568369 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-config\") pod \"neutron-db-sync-dvh5k\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.568640 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rngn\" (UniqueName: \"kubernetes.io/projected/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-kube-api-access-9rngn\") pod \"neutron-db-sync-dvh5k\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.568694 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-combined-ca-bundle\") pod \"neutron-db-sync-dvh5k\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.670247 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rngn\" (UniqueName: \"kubernetes.io/projected/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-kube-api-access-9rngn\") pod \"neutron-db-sync-dvh5k\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.670306 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-combined-ca-bundle\") pod \"neutron-db-sync-dvh5k\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.670392 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-config\") pod \"neutron-db-sync-dvh5k\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.675520 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-config\") pod \"neutron-db-sync-dvh5k\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.675919 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-combined-ca-bundle\") pod \"neutron-db-sync-dvh5k\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.689626 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rngn\" (UniqueName: \"kubernetes.io/projected/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-kube-api-access-9rngn\") pod \"neutron-db-sync-dvh5k\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:00 crc kubenswrapper[4770]: I1004 04:34:00.749204 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:01 crc kubenswrapper[4770]: I1004 04:34:01.327603 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dvh5k"] Oct 04 04:34:02 crc kubenswrapper[4770]: I1004 04:34:02.330568 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dvh5k" event={"ID":"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4","Type":"ContainerStarted","Data":"5acac5e1b13ae75d025efdeabbc5171e0281f404b3b0edfb3f5c0fbdab71671a"} Oct 04 04:34:02 crc kubenswrapper[4770]: I1004 04:34:02.331043 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dvh5k" event={"ID":"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4","Type":"ContainerStarted","Data":"629bf5db126d8521f72c0a701d2a3ba7a0fae57f6131d475dcda4b5ce554432e"} Oct 04 04:34:02 crc kubenswrapper[4770]: I1004 04:34:02.359472 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-dvh5k" podStartSLOduration=2.359429287 podStartE2EDuration="2.359429287s" podCreationTimestamp="2025-10-04 04:34:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:02.351742666 +0000 UTC m=+5453.643752378" watchObservedRunningTime="2025-10-04 04:34:02.359429287 +0000 UTC m=+5453.651439019" Oct 04 04:34:02 crc kubenswrapper[4770]: I1004 04:34:02.677168 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:34:02 crc kubenswrapper[4770]: E1004 04:34:02.677419 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:34:09 crc kubenswrapper[4770]: I1004 04:34:09.402724 4770 generic.go:334] "Generic (PLEG): container finished" podID="33d878f0-d662-4c1c-8fc4-49ccf2f7afb4" containerID="5acac5e1b13ae75d025efdeabbc5171e0281f404b3b0edfb3f5c0fbdab71671a" exitCode=0 Oct 04 04:34:09 crc kubenswrapper[4770]: I1004 04:34:09.402810 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dvh5k" event={"ID":"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4","Type":"ContainerDied","Data":"5acac5e1b13ae75d025efdeabbc5171e0281f404b3b0edfb3f5c0fbdab71671a"} Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.731741 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.853784 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-config\") pod \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.853924 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-combined-ca-bundle\") pod \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.853963 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rngn\" (UniqueName: \"kubernetes.io/projected/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-kube-api-access-9rngn\") pod \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\" (UID: \"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4\") " Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.872729 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-kube-api-access-9rngn" (OuterVolumeSpecName: "kube-api-access-9rngn") pod "33d878f0-d662-4c1c-8fc4-49ccf2f7afb4" (UID: "33d878f0-d662-4c1c-8fc4-49ccf2f7afb4"). InnerVolumeSpecName "kube-api-access-9rngn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.883443 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-config" (OuterVolumeSpecName: "config") pod "33d878f0-d662-4c1c-8fc4-49ccf2f7afb4" (UID: "33d878f0-d662-4c1c-8fc4-49ccf2f7afb4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.888488 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33d878f0-d662-4c1c-8fc4-49ccf2f7afb4" (UID: "33d878f0-d662-4c1c-8fc4-49ccf2f7afb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.974239 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.974297 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:10 crc kubenswrapper[4770]: I1004 04:34:10.974315 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rngn\" (UniqueName: \"kubernetes.io/projected/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4-kube-api-access-9rngn\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.422554 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dvh5k" event={"ID":"33d878f0-d662-4c1c-8fc4-49ccf2f7afb4","Type":"ContainerDied","Data":"629bf5db126d8521f72c0a701d2a3ba7a0fae57f6131d475dcda4b5ce554432e"} Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.422891 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="629bf5db126d8521f72c0a701d2a3ba7a0fae57f6131d475dcda4b5ce554432e" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.422627 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dvh5k" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.569694 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69f87f7665-xmlfg"] Oct 04 04:34:11 crc kubenswrapper[4770]: E1004 04:34:11.570060 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33d878f0-d662-4c1c-8fc4-49ccf2f7afb4" containerName="neutron-db-sync" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.570076 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="33d878f0-d662-4c1c-8fc4-49ccf2f7afb4" containerName="neutron-db-sync" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.571196 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="33d878f0-d662-4c1c-8fc4-49ccf2f7afb4" containerName="neutron-db-sync" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.572123 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.591927 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69f87f7665-xmlfg"] Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.693787 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-sb\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.693848 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-dns-svc\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.694047 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-nb\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.694102 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-config\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.694223 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6f4f\" (UniqueName: \"kubernetes.io/projected/b63b1f17-3ef0-4d65-8575-aea646f6c226-kube-api-access-k6f4f\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.696297 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7ddcc967fc-sj8jm"] Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.698093 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.702641 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.702876 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5ll6j" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.703138 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.721317 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7ddcc967fc-sj8jm"] Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.797353 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ea02a541-62ae-41c6-bfde-6b64f6081e8c-httpd-config\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.797503 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4mj2\" (UniqueName: \"kubernetes.io/projected/ea02a541-62ae-41c6-bfde-6b64f6081e8c-kube-api-access-l4mj2\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.797564 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ea02a541-62ae-41c6-bfde-6b64f6081e8c-config\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.797595 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-sb\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.797651 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-dns-svc\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.797702 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea02a541-62ae-41c6-bfde-6b64f6081e8c-combined-ca-bundle\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.797847 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-nb\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.798056 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-config\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.798168 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6f4f\" (UniqueName: \"kubernetes.io/projected/b63b1f17-3ef0-4d65-8575-aea646f6c226-kube-api-access-k6f4f\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.799136 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-sb\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.799315 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-nb\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.799798 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-dns-svc\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.800078 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-config\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.827113 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6f4f\" (UniqueName: \"kubernetes.io/projected/b63b1f17-3ef0-4d65-8575-aea646f6c226-kube-api-access-k6f4f\") pod \"dnsmasq-dns-69f87f7665-xmlfg\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.890458 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.899960 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ea02a541-62ae-41c6-bfde-6b64f6081e8c-httpd-config\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.900043 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4mj2\" (UniqueName: \"kubernetes.io/projected/ea02a541-62ae-41c6-bfde-6b64f6081e8c-kube-api-access-l4mj2\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.900078 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ea02a541-62ae-41c6-bfde-6b64f6081e8c-config\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.900115 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea02a541-62ae-41c6-bfde-6b64f6081e8c-combined-ca-bundle\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.915422 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ea02a541-62ae-41c6-bfde-6b64f6081e8c-httpd-config\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.918653 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea02a541-62ae-41c6-bfde-6b64f6081e8c-combined-ca-bundle\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.919093 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ea02a541-62ae-41c6-bfde-6b64f6081e8c-config\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:11 crc kubenswrapper[4770]: I1004 04:34:11.926885 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4mj2\" (UniqueName: \"kubernetes.io/projected/ea02a541-62ae-41c6-bfde-6b64f6081e8c-kube-api-access-l4mj2\") pod \"neutron-7ddcc967fc-sj8jm\" (UID: \"ea02a541-62ae-41c6-bfde-6b64f6081e8c\") " pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:12 crc kubenswrapper[4770]: I1004 04:34:12.034438 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:12 crc kubenswrapper[4770]: I1004 04:34:12.385686 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69f87f7665-xmlfg"] Oct 04 04:34:12 crc kubenswrapper[4770]: W1004 04:34:12.399666 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb63b1f17_3ef0_4d65_8575_aea646f6c226.slice/crio-be1b4ad5f6c26f3d4e44cd5a7f1acd64c3a22f17087990264a012506f14c9897 WatchSource:0}: Error finding container be1b4ad5f6c26f3d4e44cd5a7f1acd64c3a22f17087990264a012506f14c9897: Status 404 returned error can't find the container with id be1b4ad5f6c26f3d4e44cd5a7f1acd64c3a22f17087990264a012506f14c9897 Oct 04 04:34:12 crc kubenswrapper[4770]: I1004 04:34:12.433294 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" event={"ID":"b63b1f17-3ef0-4d65-8575-aea646f6c226","Type":"ContainerStarted","Data":"be1b4ad5f6c26f3d4e44cd5a7f1acd64c3a22f17087990264a012506f14c9897"} Oct 04 04:34:12 crc kubenswrapper[4770]: I1004 04:34:12.640229 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7ddcc967fc-sj8jm"] Oct 04 04:34:13 crc kubenswrapper[4770]: I1004 04:34:13.442078 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" event={"ID":"b63b1f17-3ef0-4d65-8575-aea646f6c226","Type":"ContainerStarted","Data":"f199045482713d2a6a6c1cc671e5c1e7ccec3047d15c616b1417a07f34212049"} Oct 04 04:34:13 crc kubenswrapper[4770]: I1004 04:34:13.445092 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7ddcc967fc-sj8jm" event={"ID":"ea02a541-62ae-41c6-bfde-6b64f6081e8c","Type":"ContainerStarted","Data":"d6e008dee04833db76db1d700c1f18862e8fe036ced0c4c52096ec327faf2aff"} Oct 04 04:34:13 crc kubenswrapper[4770]: I1004 04:34:13.445120 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7ddcc967fc-sj8jm" event={"ID":"ea02a541-62ae-41c6-bfde-6b64f6081e8c","Type":"ContainerStarted","Data":"2a3451e22f9a9b0973e229c53cf9b16bf9e2561f5cae5b42a3fad629199e5382"} Oct 04 04:34:14 crc kubenswrapper[4770]: I1004 04:34:14.458244 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7ddcc967fc-sj8jm" event={"ID":"ea02a541-62ae-41c6-bfde-6b64f6081e8c","Type":"ContainerStarted","Data":"973c73371903e36b4bd0e7c43954834873f7979728ef30f9c577097a437b90e0"} Oct 04 04:34:14 crc kubenswrapper[4770]: I1004 04:34:14.459223 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:14 crc kubenswrapper[4770]: I1004 04:34:14.460673 4770 generic.go:334] "Generic (PLEG): container finished" podID="b63b1f17-3ef0-4d65-8575-aea646f6c226" containerID="f199045482713d2a6a6c1cc671e5c1e7ccec3047d15c616b1417a07f34212049" exitCode=0 Oct 04 04:34:14 crc kubenswrapper[4770]: I1004 04:34:14.460701 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" event={"ID":"b63b1f17-3ef0-4d65-8575-aea646f6c226","Type":"ContainerDied","Data":"f199045482713d2a6a6c1cc671e5c1e7ccec3047d15c616b1417a07f34212049"} Oct 04 04:34:14 crc kubenswrapper[4770]: I1004 04:34:14.493068 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7ddcc967fc-sj8jm" podStartSLOduration=3.493045923 podStartE2EDuration="3.493045923s" podCreationTimestamp="2025-10-04 04:34:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:14.479673323 +0000 UTC m=+5465.771683035" watchObservedRunningTime="2025-10-04 04:34:14.493045923 +0000 UTC m=+5465.785055635" Oct 04 04:34:15 crc kubenswrapper[4770]: I1004 04:34:15.470162 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" event={"ID":"b63b1f17-3ef0-4d65-8575-aea646f6c226","Type":"ContainerStarted","Data":"672ef7ca17c64073cac265a3beb342e62dbe0e1c85567cba7a96a65a8343d071"} Oct 04 04:34:15 crc kubenswrapper[4770]: I1004 04:34:15.470528 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:15 crc kubenswrapper[4770]: I1004 04:34:15.501988 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" podStartSLOduration=4.501971351 podStartE2EDuration="4.501971351s" podCreationTimestamp="2025-10-04 04:34:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:15.489716501 +0000 UTC m=+5466.781726233" watchObservedRunningTime="2025-10-04 04:34:15.501971351 +0000 UTC m=+5466.793981073" Oct 04 04:34:15 crc kubenswrapper[4770]: I1004 04:34:15.674631 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:34:15 crc kubenswrapper[4770]: E1004 04:34:15.674858 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:34:21 crc kubenswrapper[4770]: I1004 04:34:21.892929 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:34:21 crc kubenswrapper[4770]: I1004 04:34:21.954436 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56b4bdb47c-fgzf7"] Oct 04 04:34:21 crc kubenswrapper[4770]: I1004 04:34:21.954668 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" podUID="969bbf54-1ba4-484f-84df-10521d8c2b4f" containerName="dnsmasq-dns" containerID="cri-o://898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab" gracePeriod=10 Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.434761 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.538848 4770 generic.go:334] "Generic (PLEG): container finished" podID="969bbf54-1ba4-484f-84df-10521d8c2b4f" containerID="898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab" exitCode=0 Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.538897 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" event={"ID":"969bbf54-1ba4-484f-84df-10521d8c2b4f","Type":"ContainerDied","Data":"898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab"} Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.538972 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.538996 4770 scope.go:117] "RemoveContainer" containerID="898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.538979 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56b4bdb47c-fgzf7" event={"ID":"969bbf54-1ba4-484f-84df-10521d8c2b4f","Type":"ContainerDied","Data":"9bb617c3c2e2e4f916bd41e880be5a3781f317fcb5a1abd434e363793f40b1dd"} Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.566312 4770 scope.go:117] "RemoveContainer" containerID="6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.599779 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-config\") pod \"969bbf54-1ba4-484f-84df-10521d8c2b4f\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.599878 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-nb\") pod \"969bbf54-1ba4-484f-84df-10521d8c2b4f\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.599930 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-sb\") pod \"969bbf54-1ba4-484f-84df-10521d8c2b4f\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.600049 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-dns-svc\") pod \"969bbf54-1ba4-484f-84df-10521d8c2b4f\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.600117 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fcjf\" (UniqueName: \"kubernetes.io/projected/969bbf54-1ba4-484f-84df-10521d8c2b4f-kube-api-access-7fcjf\") pod \"969bbf54-1ba4-484f-84df-10521d8c2b4f\" (UID: \"969bbf54-1ba4-484f-84df-10521d8c2b4f\") " Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.600507 4770 scope.go:117] "RemoveContainer" containerID="898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab" Oct 04 04:34:22 crc kubenswrapper[4770]: E1004 04:34:22.603275 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab\": container with ID starting with 898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab not found: ID does not exist" containerID="898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.603334 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab"} err="failed to get container status \"898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab\": rpc error: code = NotFound desc = could not find container \"898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab\": container with ID starting with 898aebe1049da001f46c200ff8af70a380c86e491cf74c1ebf39db35e4e983ab not found: ID does not exist" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.603369 4770 scope.go:117] "RemoveContainer" containerID="6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6" Oct 04 04:34:22 crc kubenswrapper[4770]: E1004 04:34:22.603857 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6\": container with ID starting with 6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6 not found: ID does not exist" containerID="6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.603910 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6"} err="failed to get container status \"6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6\": rpc error: code = NotFound desc = could not find container \"6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6\": container with ID starting with 6c003c995142dd4eabc7e814876b1c5686e280e63b3b26d165cc0fabe3ae2be6 not found: ID does not exist" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.607265 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/969bbf54-1ba4-484f-84df-10521d8c2b4f-kube-api-access-7fcjf" (OuterVolumeSpecName: "kube-api-access-7fcjf") pod "969bbf54-1ba4-484f-84df-10521d8c2b4f" (UID: "969bbf54-1ba4-484f-84df-10521d8c2b4f"). InnerVolumeSpecName "kube-api-access-7fcjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.659347 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-config" (OuterVolumeSpecName: "config") pod "969bbf54-1ba4-484f-84df-10521d8c2b4f" (UID: "969bbf54-1ba4-484f-84df-10521d8c2b4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.661172 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "969bbf54-1ba4-484f-84df-10521d8c2b4f" (UID: "969bbf54-1ba4-484f-84df-10521d8c2b4f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.668770 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "969bbf54-1ba4-484f-84df-10521d8c2b4f" (UID: "969bbf54-1ba4-484f-84df-10521d8c2b4f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.672631 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "969bbf54-1ba4-484f-84df-10521d8c2b4f" (UID: "969bbf54-1ba4-484f-84df-10521d8c2b4f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.702181 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fcjf\" (UniqueName: \"kubernetes.io/projected/969bbf54-1ba4-484f-84df-10521d8c2b4f-kube-api-access-7fcjf\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.702212 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.702222 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.702233 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.702241 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/969bbf54-1ba4-484f-84df-10521d8c2b4f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.869576 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56b4bdb47c-fgzf7"] Oct 04 04:34:22 crc kubenswrapper[4770]: I1004 04:34:22.875700 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56b4bdb47c-fgzf7"] Oct 04 04:34:23 crc kubenswrapper[4770]: I1004 04:34:23.683765 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="969bbf54-1ba4-484f-84df-10521d8c2b4f" path="/var/lib/kubelet/pods/969bbf54-1ba4-484f-84df-10521d8c2b4f/volumes" Oct 04 04:34:27 crc kubenswrapper[4770]: I1004 04:34:27.674207 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:34:27 crc kubenswrapper[4770]: E1004 04:34:27.674905 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:34:41 crc kubenswrapper[4770]: I1004 04:34:41.674055 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:34:41 crc kubenswrapper[4770]: E1004 04:34:41.674752 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:34:42 crc kubenswrapper[4770]: I1004 04:34:42.042910 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7ddcc967fc-sj8jm" Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.453055 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-9kk82"] Oct 04 04:34:52 crc kubenswrapper[4770]: E1004 04:34:52.453980 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969bbf54-1ba4-484f-84df-10521d8c2b4f" containerName="dnsmasq-dns" Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.454000 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="969bbf54-1ba4-484f-84df-10521d8c2b4f" containerName="dnsmasq-dns" Oct 04 04:34:52 crc kubenswrapper[4770]: E1004 04:34:52.454044 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969bbf54-1ba4-484f-84df-10521d8c2b4f" containerName="init" Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.454052 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="969bbf54-1ba4-484f-84df-10521d8c2b4f" containerName="init" Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.454273 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="969bbf54-1ba4-484f-84df-10521d8c2b4f" containerName="dnsmasq-dns" Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.454949 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9kk82" Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.462161 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9kk82"] Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.497790 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hwbp\" (UniqueName: \"kubernetes.io/projected/905a1811-a6a0-4172-8d4b-1a34bf20865a-kube-api-access-7hwbp\") pod \"glance-db-create-9kk82\" (UID: \"905a1811-a6a0-4172-8d4b-1a34bf20865a\") " pod="openstack/glance-db-create-9kk82" Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.599059 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hwbp\" (UniqueName: \"kubernetes.io/projected/905a1811-a6a0-4172-8d4b-1a34bf20865a-kube-api-access-7hwbp\") pod \"glance-db-create-9kk82\" (UID: \"905a1811-a6a0-4172-8d4b-1a34bf20865a\") " pod="openstack/glance-db-create-9kk82" Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.637939 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hwbp\" (UniqueName: \"kubernetes.io/projected/905a1811-a6a0-4172-8d4b-1a34bf20865a-kube-api-access-7hwbp\") pod \"glance-db-create-9kk82\" (UID: \"905a1811-a6a0-4172-8d4b-1a34bf20865a\") " pod="openstack/glance-db-create-9kk82" Oct 04 04:34:52 crc kubenswrapper[4770]: I1004 04:34:52.781552 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9kk82" Oct 04 04:34:53 crc kubenswrapper[4770]: I1004 04:34:53.269450 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9kk82"] Oct 04 04:34:53 crc kubenswrapper[4770]: I1004 04:34:53.828124 4770 generic.go:334] "Generic (PLEG): container finished" podID="905a1811-a6a0-4172-8d4b-1a34bf20865a" containerID="5fc058ab5405739055c9607f76969dd47f409c9f692576783792c47157490033" exitCode=0 Oct 04 04:34:53 crc kubenswrapper[4770]: I1004 04:34:53.828176 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9kk82" event={"ID":"905a1811-a6a0-4172-8d4b-1a34bf20865a","Type":"ContainerDied","Data":"5fc058ab5405739055c9607f76969dd47f409c9f692576783792c47157490033"} Oct 04 04:34:53 crc kubenswrapper[4770]: I1004 04:34:53.828420 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9kk82" event={"ID":"905a1811-a6a0-4172-8d4b-1a34bf20865a","Type":"ContainerStarted","Data":"1f40c3533b3ab5049474663b135935b892bada4b521cc78de583da7d46535bf0"} Oct 04 04:34:55 crc kubenswrapper[4770]: I1004 04:34:55.156935 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9kk82" Oct 04 04:34:55 crc kubenswrapper[4770]: I1004 04:34:55.353819 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hwbp\" (UniqueName: \"kubernetes.io/projected/905a1811-a6a0-4172-8d4b-1a34bf20865a-kube-api-access-7hwbp\") pod \"905a1811-a6a0-4172-8d4b-1a34bf20865a\" (UID: \"905a1811-a6a0-4172-8d4b-1a34bf20865a\") " Oct 04 04:34:55 crc kubenswrapper[4770]: I1004 04:34:55.359853 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/905a1811-a6a0-4172-8d4b-1a34bf20865a-kube-api-access-7hwbp" (OuterVolumeSpecName: "kube-api-access-7hwbp") pod "905a1811-a6a0-4172-8d4b-1a34bf20865a" (UID: "905a1811-a6a0-4172-8d4b-1a34bf20865a"). InnerVolumeSpecName "kube-api-access-7hwbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:34:55 crc kubenswrapper[4770]: I1004 04:34:55.455949 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hwbp\" (UniqueName: \"kubernetes.io/projected/905a1811-a6a0-4172-8d4b-1a34bf20865a-kube-api-access-7hwbp\") on node \"crc\" DevicePath \"\"" Oct 04 04:34:55 crc kubenswrapper[4770]: I1004 04:34:55.849035 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9kk82" event={"ID":"905a1811-a6a0-4172-8d4b-1a34bf20865a","Type":"ContainerDied","Data":"1f40c3533b3ab5049474663b135935b892bada4b521cc78de583da7d46535bf0"} Oct 04 04:34:55 crc kubenswrapper[4770]: I1004 04:34:55.849421 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f40c3533b3ab5049474663b135935b892bada4b521cc78de583da7d46535bf0" Oct 04 04:34:55 crc kubenswrapper[4770]: I1004 04:34:55.849102 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9kk82" Oct 04 04:34:56 crc kubenswrapper[4770]: I1004 04:34:56.674469 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:34:56 crc kubenswrapper[4770]: E1004 04:34:56.674727 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.534764 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-685f-account-create-7zfxz"] Oct 04 04:35:02 crc kubenswrapper[4770]: E1004 04:35:02.535772 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="905a1811-a6a0-4172-8d4b-1a34bf20865a" containerName="mariadb-database-create" Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.535789 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="905a1811-a6a0-4172-8d4b-1a34bf20865a" containerName="mariadb-database-create" Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.536041 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="905a1811-a6a0-4172-8d4b-1a34bf20865a" containerName="mariadb-database-create" Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.536854 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-685f-account-create-7zfxz" Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.539309 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.546985 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-685f-account-create-7zfxz"] Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.586824 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d95t5\" (UniqueName: \"kubernetes.io/projected/e0846dbf-391e-4431-923e-1bf4d138be4b-kube-api-access-d95t5\") pod \"glance-685f-account-create-7zfxz\" (UID: \"e0846dbf-391e-4431-923e-1bf4d138be4b\") " pod="openstack/glance-685f-account-create-7zfxz" Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.688650 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d95t5\" (UniqueName: \"kubernetes.io/projected/e0846dbf-391e-4431-923e-1bf4d138be4b-kube-api-access-d95t5\") pod \"glance-685f-account-create-7zfxz\" (UID: \"e0846dbf-391e-4431-923e-1bf4d138be4b\") " pod="openstack/glance-685f-account-create-7zfxz" Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.707209 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d95t5\" (UniqueName: \"kubernetes.io/projected/e0846dbf-391e-4431-923e-1bf4d138be4b-kube-api-access-d95t5\") pod \"glance-685f-account-create-7zfxz\" (UID: \"e0846dbf-391e-4431-923e-1bf4d138be4b\") " pod="openstack/glance-685f-account-create-7zfxz" Oct 04 04:35:02 crc kubenswrapper[4770]: I1004 04:35:02.859303 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-685f-account-create-7zfxz" Oct 04 04:35:03 crc kubenswrapper[4770]: I1004 04:35:03.359111 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-685f-account-create-7zfxz"] Oct 04 04:35:03 crc kubenswrapper[4770]: I1004 04:35:03.932446 4770 generic.go:334] "Generic (PLEG): container finished" podID="e0846dbf-391e-4431-923e-1bf4d138be4b" containerID="adbc6ea3b629a932317e2bc40100cb81459d77694b33e6802b7d653870f344ec" exitCode=0 Oct 04 04:35:03 crc kubenswrapper[4770]: I1004 04:35:03.932540 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-685f-account-create-7zfxz" event={"ID":"e0846dbf-391e-4431-923e-1bf4d138be4b","Type":"ContainerDied","Data":"adbc6ea3b629a932317e2bc40100cb81459d77694b33e6802b7d653870f344ec"} Oct 04 04:35:03 crc kubenswrapper[4770]: I1004 04:35:03.932778 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-685f-account-create-7zfxz" event={"ID":"e0846dbf-391e-4431-923e-1bf4d138be4b","Type":"ContainerStarted","Data":"79b729b06c2a6cb653c134a789de9e16bd480d3db1e913d21500ad7fb6d6846e"} Oct 04 04:35:05 crc kubenswrapper[4770]: I1004 04:35:05.242189 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-685f-account-create-7zfxz" Oct 04 04:35:05 crc kubenswrapper[4770]: I1004 04:35:05.433081 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d95t5\" (UniqueName: \"kubernetes.io/projected/e0846dbf-391e-4431-923e-1bf4d138be4b-kube-api-access-d95t5\") pod \"e0846dbf-391e-4431-923e-1bf4d138be4b\" (UID: \"e0846dbf-391e-4431-923e-1bf4d138be4b\") " Oct 04 04:35:05 crc kubenswrapper[4770]: I1004 04:35:05.441236 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0846dbf-391e-4431-923e-1bf4d138be4b-kube-api-access-d95t5" (OuterVolumeSpecName: "kube-api-access-d95t5") pod "e0846dbf-391e-4431-923e-1bf4d138be4b" (UID: "e0846dbf-391e-4431-923e-1bf4d138be4b"). InnerVolumeSpecName "kube-api-access-d95t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:05 crc kubenswrapper[4770]: I1004 04:35:05.536180 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d95t5\" (UniqueName: \"kubernetes.io/projected/e0846dbf-391e-4431-923e-1bf4d138be4b-kube-api-access-d95t5\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:05 crc kubenswrapper[4770]: I1004 04:35:05.952331 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-685f-account-create-7zfxz" event={"ID":"e0846dbf-391e-4431-923e-1bf4d138be4b","Type":"ContainerDied","Data":"79b729b06c2a6cb653c134a789de9e16bd480d3db1e913d21500ad7fb6d6846e"} Oct 04 04:35:05 crc kubenswrapper[4770]: I1004 04:35:05.952603 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79b729b06c2a6cb653c134a789de9e16bd480d3db1e913d21500ad7fb6d6846e" Oct 04 04:35:05 crc kubenswrapper[4770]: I1004 04:35:05.952390 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-685f-account-create-7zfxz" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.608171 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-t5797"] Oct 04 04:35:07 crc kubenswrapper[4770]: E1004 04:35:07.614127 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0846dbf-391e-4431-923e-1bf4d138be4b" containerName="mariadb-account-create" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.614448 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0846dbf-391e-4431-923e-1bf4d138be4b" containerName="mariadb-account-create" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.614723 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0846dbf-391e-4431-923e-1bf4d138be4b" containerName="mariadb-account-create" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.615544 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.620355 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.621218 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mm74r" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.621837 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-t5797"] Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.773982 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-combined-ca-bundle\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.774312 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-db-sync-config-data\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.774365 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-config-data\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.774482 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x7mp\" (UniqueName: \"kubernetes.io/projected/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-kube-api-access-5x7mp\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.876689 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-combined-ca-bundle\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.876831 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-db-sync-config-data\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.876863 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-config-data\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.876924 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x7mp\" (UniqueName: \"kubernetes.io/projected/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-kube-api-access-5x7mp\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.883468 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-config-data\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.893908 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-db-sync-config-data\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.896928 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-combined-ca-bundle\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.903614 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x7mp\" (UniqueName: \"kubernetes.io/projected/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-kube-api-access-5x7mp\") pod \"glance-db-sync-t5797\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " pod="openstack/glance-db-sync-t5797" Oct 04 04:35:07 crc kubenswrapper[4770]: I1004 04:35:07.937388 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-t5797" Oct 04 04:35:08 crc kubenswrapper[4770]: I1004 04:35:08.504659 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-t5797"] Oct 04 04:35:08 crc kubenswrapper[4770]: I1004 04:35:08.982023 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-t5797" event={"ID":"7bf96ca3-0b4a-489a-864a-d0299e1db7cc","Type":"ContainerStarted","Data":"05ead101907547ff10a4d21ffbc8b64fb6d542b655c30b165f50e27890986eec"} Oct 04 04:35:09 crc kubenswrapper[4770]: I1004 04:35:09.712497 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:35:09 crc kubenswrapper[4770]: E1004 04:35:09.713082 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:35:09 crc kubenswrapper[4770]: I1004 04:35:09.996776 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-t5797" event={"ID":"7bf96ca3-0b4a-489a-864a-d0299e1db7cc","Type":"ContainerStarted","Data":"44bbe576224173f3afb1a8b498626bffcd059b8b3391a2d339cfb57981fc7ce0"} Oct 04 04:35:13 crc kubenswrapper[4770]: I1004 04:35:13.027712 4770 generic.go:334] "Generic (PLEG): container finished" podID="7bf96ca3-0b4a-489a-864a-d0299e1db7cc" containerID="44bbe576224173f3afb1a8b498626bffcd059b8b3391a2d339cfb57981fc7ce0" exitCode=0 Oct 04 04:35:13 crc kubenswrapper[4770]: I1004 04:35:13.027938 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-t5797" event={"ID":"7bf96ca3-0b4a-489a-864a-d0299e1db7cc","Type":"ContainerDied","Data":"44bbe576224173f3afb1a8b498626bffcd059b8b3391a2d339cfb57981fc7ce0"} Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.516642 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-t5797" Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.692313 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x7mp\" (UniqueName: \"kubernetes.io/projected/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-kube-api-access-5x7mp\") pod \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.692437 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-config-data\") pod \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.692572 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-db-sync-config-data\") pod \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.692736 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-combined-ca-bundle\") pod \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\" (UID: \"7bf96ca3-0b4a-489a-864a-d0299e1db7cc\") " Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.701617 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-kube-api-access-5x7mp" (OuterVolumeSpecName: "kube-api-access-5x7mp") pod "7bf96ca3-0b4a-489a-864a-d0299e1db7cc" (UID: "7bf96ca3-0b4a-489a-864a-d0299e1db7cc"). InnerVolumeSpecName "kube-api-access-5x7mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.701623 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7bf96ca3-0b4a-489a-864a-d0299e1db7cc" (UID: "7bf96ca3-0b4a-489a-864a-d0299e1db7cc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.744464 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bf96ca3-0b4a-489a-864a-d0299e1db7cc" (UID: "7bf96ca3-0b4a-489a-864a-d0299e1db7cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.769885 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-config-data" (OuterVolumeSpecName: "config-data") pod "7bf96ca3-0b4a-489a-864a-d0299e1db7cc" (UID: "7bf96ca3-0b4a-489a-864a-d0299e1db7cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.796637 4770 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.796943 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.797230 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x7mp\" (UniqueName: \"kubernetes.io/projected/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-kube-api-access-5x7mp\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:14 crc kubenswrapper[4770]: I1004 04:35:14.797383 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf96ca3-0b4a-489a-864a-d0299e1db7cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.050537 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-t5797" event={"ID":"7bf96ca3-0b4a-489a-864a-d0299e1db7cc","Type":"ContainerDied","Data":"05ead101907547ff10a4d21ffbc8b64fb6d542b655c30b165f50e27890986eec"} Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.050963 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05ead101907547ff10a4d21ffbc8b64fb6d542b655c30b165f50e27890986eec" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.051070 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-t5797" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.359346 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:35:15 crc kubenswrapper[4770]: E1004 04:35:15.359819 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf96ca3-0b4a-489a-864a-d0299e1db7cc" containerName="glance-db-sync" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.359846 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf96ca3-0b4a-489a-864a-d0299e1db7cc" containerName="glance-db-sync" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.360123 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bf96ca3-0b4a-489a-864a-d0299e1db7cc" containerName="glance-db-sync" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.361500 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.367304 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.367361 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.367313 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mm74r" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.367577 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.447189 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.516322 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d4d6fcd69-chhqb"] Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.518491 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526339 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-config-data\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526384 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6bf9\" (UniqueName: \"kubernetes.io/projected/0684e72c-e2a5-4982-8f16-8cc8b10a152a-kube-api-access-v6bf9\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526405 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btwjf\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-kube-api-access-btwjf\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526455 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-logs\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526490 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526523 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-config\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526544 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-sb\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526560 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-nb\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526583 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526600 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-scripts\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526618 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-ceph\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.526638 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-dns-svc\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.542140 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d4d6fcd69-chhqb"] Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.553395 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.554796 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.556745 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.576027 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.628853 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.628904 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-logs\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.628955 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.628984 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629034 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629077 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-config\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629107 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-sb\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629124 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-nb\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629150 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629170 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-scripts\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629186 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629208 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-ceph\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629226 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-dns-svc\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629247 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjxsf\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-kube-api-access-sjxsf\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629273 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-config-data\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629292 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6bf9\" (UniqueName: \"kubernetes.io/projected/0684e72c-e2a5-4982-8f16-8cc8b10a152a-kube-api-access-v6bf9\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629308 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btwjf\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-kube-api-access-btwjf\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629344 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.629363 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.630343 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-logs\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.631286 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.631594 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-nb\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.632250 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-config\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.632782 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-sb\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.633174 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-dns-svc\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.634949 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-ceph\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.637315 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-config-data\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.648937 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.652726 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btwjf\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-kube-api-access-btwjf\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.655445 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6bf9\" (UniqueName: \"kubernetes.io/projected/0684e72c-e2a5-4982-8f16-8cc8b10a152a-kube-api-access-v6bf9\") pod \"dnsmasq-dns-7d4d6fcd69-chhqb\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.656259 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-scripts\") pod \"glance-default-external-api-0\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.731872 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.731937 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.731996 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.732056 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjxsf\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-kube-api-access-sjxsf\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.732102 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.732130 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.732179 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.732380 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.732849 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.735708 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.736329 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.736549 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.740956 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.750644 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.752339 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjxsf\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-kube-api-access-sjxsf\") pod \"glance-default-internal-api-0\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.861885 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:15 crc kubenswrapper[4770]: I1004 04:35:15.886336 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:16 crc kubenswrapper[4770]: I1004 04:35:16.298312 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:35:16 crc kubenswrapper[4770]: I1004 04:35:16.344070 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d4d6fcd69-chhqb"] Oct 04 04:35:16 crc kubenswrapper[4770]: I1004 04:35:16.525094 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:35:16 crc kubenswrapper[4770]: I1004 04:35:16.909753 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:35:17 crc kubenswrapper[4770]: I1004 04:35:17.076352 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88594c99-e1c5-4f9f-b38b-fb125afbd5dc","Type":"ContainerStarted","Data":"6847237d5dd900bb61cf5d7760e835f2b09a81e03d55a73e5323b80cb74b8b57"} Oct 04 04:35:17 crc kubenswrapper[4770]: I1004 04:35:17.078017 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a6e8889-c360-4e35-be4f-48160edcbb9f","Type":"ContainerStarted","Data":"57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983"} Oct 04 04:35:17 crc kubenswrapper[4770]: I1004 04:35:17.078054 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a6e8889-c360-4e35-be4f-48160edcbb9f","Type":"ContainerStarted","Data":"4113c5ad8ceab17e50e125e910b10d766537d9cc1df9356073b7973543962345"} Oct 04 04:35:17 crc kubenswrapper[4770]: I1004 04:35:17.079701 4770 generic.go:334] "Generic (PLEG): container finished" podID="0684e72c-e2a5-4982-8f16-8cc8b10a152a" containerID="a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad" exitCode=0 Oct 04 04:35:17 crc kubenswrapper[4770]: I1004 04:35:17.079728 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" event={"ID":"0684e72c-e2a5-4982-8f16-8cc8b10a152a","Type":"ContainerDied","Data":"a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad"} Oct 04 04:35:17 crc kubenswrapper[4770]: I1004 04:35:17.079745 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" event={"ID":"0684e72c-e2a5-4982-8f16-8cc8b10a152a","Type":"ContainerStarted","Data":"5b805fbd6c67e3cff5f09c27975418e2b8c9758836666df7b6412a4c68435d7f"} Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.093229 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a6e8889-c360-4e35-be4f-48160edcbb9f","Type":"ContainerStarted","Data":"a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af"} Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.093297 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerName="glance-log" containerID="cri-o://57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983" gracePeriod=30 Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.093341 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerName="glance-httpd" containerID="cri-o://a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af" gracePeriod=30 Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.096183 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" event={"ID":"0684e72c-e2a5-4982-8f16-8cc8b10a152a","Type":"ContainerStarted","Data":"8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d"} Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.096497 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.102481 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88594c99-e1c5-4f9f-b38b-fb125afbd5dc","Type":"ContainerStarted","Data":"9c34217414250802f574c36e9b60141304f8d86276f493b08fa496a1fb3bda73"} Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.102512 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88594c99-e1c5-4f9f-b38b-fb125afbd5dc","Type":"ContainerStarted","Data":"fca6b372a2602f442b8c6a814b63633b38c48169b607e18757b127aa120b4c92"} Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.114845 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.114823168 podStartE2EDuration="3.114823168s" podCreationTimestamp="2025-10-04 04:35:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:18.113414782 +0000 UTC m=+5529.405424504" watchObservedRunningTime="2025-10-04 04:35:18.114823168 +0000 UTC m=+5529.406832900" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.162258 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.162239638 podStartE2EDuration="3.162239638s" podCreationTimestamp="2025-10-04 04:35:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:18.157898804 +0000 UTC m=+5529.449908516" watchObservedRunningTime="2025-10-04 04:35:18.162239638 +0000 UTC m=+5529.454249350" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.184336 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" podStartSLOduration=3.184312275 podStartE2EDuration="3.184312275s" podCreationTimestamp="2025-10-04 04:35:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:18.182380734 +0000 UTC m=+5529.474390466" watchObservedRunningTime="2025-10-04 04:35:18.184312275 +0000 UTC m=+5529.476322007" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.789411 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.906884 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-scripts\") pod \"9a6e8889-c360-4e35-be4f-48160edcbb9f\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.906976 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-httpd-run\") pod \"9a6e8889-c360-4e35-be4f-48160edcbb9f\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.906994 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-ceph\") pod \"9a6e8889-c360-4e35-be4f-48160edcbb9f\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.907071 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-logs\") pod \"9a6e8889-c360-4e35-be4f-48160edcbb9f\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.907166 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-combined-ca-bundle\") pod \"9a6e8889-c360-4e35-be4f-48160edcbb9f\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.907199 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-config-data\") pod \"9a6e8889-c360-4e35-be4f-48160edcbb9f\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.907276 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btwjf\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-kube-api-access-btwjf\") pod \"9a6e8889-c360-4e35-be4f-48160edcbb9f\" (UID: \"9a6e8889-c360-4e35-be4f-48160edcbb9f\") " Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.907594 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-logs" (OuterVolumeSpecName: "logs") pod "9a6e8889-c360-4e35-be4f-48160edcbb9f" (UID: "9a6e8889-c360-4e35-be4f-48160edcbb9f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.907686 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.907503 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9a6e8889-c360-4e35-be4f-48160edcbb9f" (UID: "9a6e8889-c360-4e35-be4f-48160edcbb9f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.918094 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-scripts" (OuterVolumeSpecName: "scripts") pod "9a6e8889-c360-4e35-be4f-48160edcbb9f" (UID: "9a6e8889-c360-4e35-be4f-48160edcbb9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.918182 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-ceph" (OuterVolumeSpecName: "ceph") pod "9a6e8889-c360-4e35-be4f-48160edcbb9f" (UID: "9a6e8889-c360-4e35-be4f-48160edcbb9f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.918242 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-kube-api-access-btwjf" (OuterVolumeSpecName: "kube-api-access-btwjf") pod "9a6e8889-c360-4e35-be4f-48160edcbb9f" (UID: "9a6e8889-c360-4e35-be4f-48160edcbb9f"). InnerVolumeSpecName "kube-api-access-btwjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.944469 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a6e8889-c360-4e35-be4f-48160edcbb9f" (UID: "9a6e8889-c360-4e35-be4f-48160edcbb9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:18 crc kubenswrapper[4770]: I1004 04:35:18.974184 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-config-data" (OuterVolumeSpecName: "config-data") pod "9a6e8889-c360-4e35-be4f-48160edcbb9f" (UID: "9a6e8889-c360-4e35-be4f-48160edcbb9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.010432 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btwjf\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-kube-api-access-btwjf\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.010563 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.010633 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6e8889-c360-4e35-be4f-48160edcbb9f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.010694 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9a6e8889-c360-4e35-be4f-48160edcbb9f-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.010748 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.010802 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6e8889-c360-4e35-be4f-48160edcbb9f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.116215 4770 generic.go:334] "Generic (PLEG): container finished" podID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerID="a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af" exitCode=0 Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.116248 4770 generic.go:334] "Generic (PLEG): container finished" podID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerID="57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983" exitCode=143 Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.116299 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.117964 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a6e8889-c360-4e35-be4f-48160edcbb9f","Type":"ContainerDied","Data":"a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af"} Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.118299 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a6e8889-c360-4e35-be4f-48160edcbb9f","Type":"ContainerDied","Data":"57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983"} Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.118449 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a6e8889-c360-4e35-be4f-48160edcbb9f","Type":"ContainerDied","Data":"4113c5ad8ceab17e50e125e910b10d766537d9cc1df9356073b7973543962345"} Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.118372 4770 scope.go:117] "RemoveContainer" containerID="a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.150426 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.175639 4770 scope.go:117] "RemoveContainer" containerID="57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.178220 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.195107 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:35:19 crc kubenswrapper[4770]: E1004 04:35:19.195836 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerName="glance-log" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.195860 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerName="glance-log" Oct 04 04:35:19 crc kubenswrapper[4770]: E1004 04:35:19.195889 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerName="glance-httpd" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.195899 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerName="glance-httpd" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.196129 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerName="glance-httpd" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.196161 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a6e8889-c360-4e35-be4f-48160edcbb9f" containerName="glance-log" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.197419 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.200082 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.202335 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.222123 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-scripts\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.222216 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-logs\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.222305 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx8qt\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-kube-api-access-fx8qt\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.222340 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-config-data\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.222365 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.222394 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.222417 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-ceph\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.230185 4770 scope.go:117] "RemoveContainer" containerID="a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af" Oct 04 04:35:19 crc kubenswrapper[4770]: E1004 04:35:19.230726 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af\": container with ID starting with a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af not found: ID does not exist" containerID="a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.230854 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af"} err="failed to get container status \"a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af\": rpc error: code = NotFound desc = could not find container \"a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af\": container with ID starting with a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af not found: ID does not exist" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.230992 4770 scope.go:117] "RemoveContainer" containerID="57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983" Oct 04 04:35:19 crc kubenswrapper[4770]: E1004 04:35:19.231709 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983\": container with ID starting with 57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983 not found: ID does not exist" containerID="57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.231754 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983"} err="failed to get container status \"57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983\": rpc error: code = NotFound desc = could not find container \"57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983\": container with ID starting with 57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983 not found: ID does not exist" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.231786 4770 scope.go:117] "RemoveContainer" containerID="a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.232096 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af"} err="failed to get container status \"a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af\": rpc error: code = NotFound desc = could not find container \"a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af\": container with ID starting with a8743f6b45be6f2e209744119d4fc9af514d1dd2602112b0ba3615e5578c59af not found: ID does not exist" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.232127 4770 scope.go:117] "RemoveContainer" containerID="57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.232381 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983"} err="failed to get container status \"57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983\": rpc error: code = NotFound desc = could not find container \"57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983\": container with ID starting with 57671307de06a1cb996d9f7a961afd9a31e99c899b887dc4eef5c48feed31983 not found: ID does not exist" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.324495 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-scripts\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.324804 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-logs\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.324843 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx8qt\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-kube-api-access-fx8qt\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.324872 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-config-data\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.324891 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.324918 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.324937 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-ceph\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.325550 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.325635 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-logs\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.328592 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-scripts\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.329739 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.329865 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-ceph\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.330550 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-config-data\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.341247 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx8qt\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-kube-api-access-fx8qt\") pod \"glance-default-external-api-0\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.427692 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.514587 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:35:19 crc kubenswrapper[4770]: I1004 04:35:19.692305 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a6e8889-c360-4e35-be4f-48160edcbb9f" path="/var/lib/kubelet/pods/9a6e8889-c360-4e35-be4f-48160edcbb9f/volumes" Oct 04 04:35:20 crc kubenswrapper[4770]: I1004 04:35:20.095510 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:35:20 crc kubenswrapper[4770]: W1004 04:35:20.099435 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61d14385_5018_4129_9421_a92154491a56.slice/crio-a0ff65de50f98b756aac9bba08dc7a80022391746c229dfaa27788007baa6c27 WatchSource:0}: Error finding container a0ff65de50f98b756aac9bba08dc7a80022391746c229dfaa27788007baa6c27: Status 404 returned error can't find the container with id a0ff65de50f98b756aac9bba08dc7a80022391746c229dfaa27788007baa6c27 Oct 04 04:35:20 crc kubenswrapper[4770]: I1004 04:35:20.127865 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerName="glance-log" containerID="cri-o://fca6b372a2602f442b8c6a814b63633b38c48169b607e18757b127aa120b4c92" gracePeriod=30 Oct 04 04:35:20 crc kubenswrapper[4770]: I1004 04:35:20.128128 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"61d14385-5018-4129-9421-a92154491a56","Type":"ContainerStarted","Data":"a0ff65de50f98b756aac9bba08dc7a80022391746c229dfaa27788007baa6c27"} Oct 04 04:35:20 crc kubenswrapper[4770]: I1004 04:35:20.128361 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerName="glance-httpd" containerID="cri-o://9c34217414250802f574c36e9b60141304f8d86276f493b08fa496a1fb3bda73" gracePeriod=30 Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.150538 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"61d14385-5018-4129-9421-a92154491a56","Type":"ContainerStarted","Data":"9cbadeb8d48ac7bab37fe817e22037ac25e1ca21b04f53adb1faf1de0c47076a"} Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.157304 4770 generic.go:334] "Generic (PLEG): container finished" podID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerID="9c34217414250802f574c36e9b60141304f8d86276f493b08fa496a1fb3bda73" exitCode=0 Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.157340 4770 generic.go:334] "Generic (PLEG): container finished" podID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerID="fca6b372a2602f442b8c6a814b63633b38c48169b607e18757b127aa120b4c92" exitCode=143 Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.157365 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88594c99-e1c5-4f9f-b38b-fb125afbd5dc","Type":"ContainerDied","Data":"9c34217414250802f574c36e9b60141304f8d86276f493b08fa496a1fb3bda73"} Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.157396 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88594c99-e1c5-4f9f-b38b-fb125afbd5dc","Type":"ContainerDied","Data":"fca6b372a2602f442b8c6a814b63633b38c48169b607e18757b127aa120b4c92"} Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.276106 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.380199 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-logs\") pod \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.380560 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-ceph\") pod \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.380589 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-httpd-run\") pod \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.380613 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-scripts\") pod \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.380667 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjxsf\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-kube-api-access-sjxsf\") pod \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.380789 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-logs" (OuterVolumeSpecName: "logs") pod "88594c99-e1c5-4f9f-b38b-fb125afbd5dc" (UID: "88594c99-e1c5-4f9f-b38b-fb125afbd5dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.381030 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "88594c99-e1c5-4f9f-b38b-fb125afbd5dc" (UID: "88594c99-e1c5-4f9f-b38b-fb125afbd5dc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.381322 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-combined-ca-bundle\") pod \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.381821 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-config-data\") pod \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\" (UID: \"88594c99-e1c5-4f9f-b38b-fb125afbd5dc\") " Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.382514 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.382623 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.387620 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-kube-api-access-sjxsf" (OuterVolumeSpecName: "kube-api-access-sjxsf") pod "88594c99-e1c5-4f9f-b38b-fb125afbd5dc" (UID: "88594c99-e1c5-4f9f-b38b-fb125afbd5dc"). InnerVolumeSpecName "kube-api-access-sjxsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.387712 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-ceph" (OuterVolumeSpecName: "ceph") pod "88594c99-e1c5-4f9f-b38b-fb125afbd5dc" (UID: "88594c99-e1c5-4f9f-b38b-fb125afbd5dc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.398229 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-scripts" (OuterVolumeSpecName: "scripts") pod "88594c99-e1c5-4f9f-b38b-fb125afbd5dc" (UID: "88594c99-e1c5-4f9f-b38b-fb125afbd5dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.409432 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88594c99-e1c5-4f9f-b38b-fb125afbd5dc" (UID: "88594c99-e1c5-4f9f-b38b-fb125afbd5dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.430536 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-config-data" (OuterVolumeSpecName: "config-data") pod "88594c99-e1c5-4f9f-b38b-fb125afbd5dc" (UID: "88594c99-e1c5-4f9f-b38b-fb125afbd5dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.483738 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.483777 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.483786 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.483796 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjxsf\" (UniqueName: \"kubernetes.io/projected/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-kube-api-access-sjxsf\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:21 crc kubenswrapper[4770]: I1004 04:35:21.483806 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88594c99-e1c5-4f9f-b38b-fb125afbd5dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.166878 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"61d14385-5018-4129-9421-a92154491a56","Type":"ContainerStarted","Data":"4c463a78a6efad8325eb9f8ea8fcef2c9fb12b91d453d406495e3499bb0b2c38"} Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.168655 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88594c99-e1c5-4f9f-b38b-fb125afbd5dc","Type":"ContainerDied","Data":"6847237d5dd900bb61cf5d7760e835f2b09a81e03d55a73e5323b80cb74b8b57"} Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.168710 4770 scope.go:117] "RemoveContainer" containerID="9c34217414250802f574c36e9b60141304f8d86276f493b08fa496a1fb3bda73" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.168756 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.187990 4770 scope.go:117] "RemoveContainer" containerID="fca6b372a2602f442b8c6a814b63633b38c48169b607e18757b127aa120b4c92" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.214556 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.214536368 podStartE2EDuration="3.214536368s" podCreationTimestamp="2025-10-04 04:35:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:22.188058896 +0000 UTC m=+5533.480068628" watchObservedRunningTime="2025-10-04 04:35:22.214536368 +0000 UTC m=+5533.506546080" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.216266 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.225531 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.234425 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:35:22 crc kubenswrapper[4770]: E1004 04:35:22.234935 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerName="glance-httpd" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.234955 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerName="glance-httpd" Oct 04 04:35:22 crc kubenswrapper[4770]: E1004 04:35:22.234984 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerName="glance-log" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.234991 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerName="glance-log" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.235231 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerName="glance-log" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.235266 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" containerName="glance-httpd" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.236193 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.241434 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.250640 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.400168 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.400293 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.400367 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.400574 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.400713 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.400774 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.400833 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sppms\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-kube-api-access-sppms\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.502484 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.502748 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.502824 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.502882 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.502929 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sppms\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-kube-api-access-sppms\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.503429 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.503668 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.504377 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.504705 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.511678 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.511910 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.512048 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.512419 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.529275 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sppms\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-kube-api-access-sppms\") pod \"glance-default-internal-api-0\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:35:22 crc kubenswrapper[4770]: I1004 04:35:22.560862 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:23 crc kubenswrapper[4770]: I1004 04:35:23.250788 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:35:23 crc kubenswrapper[4770]: W1004 04:35:23.257956 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7783d14_4f52_47fa_8bb9_5a99cb5f7192.slice/crio-40d87c7a01f3bd728ef41dee6e00a6fa7d102d7be1e78050c4cbf3a1e7d0a378 WatchSource:0}: Error finding container 40d87c7a01f3bd728ef41dee6e00a6fa7d102d7be1e78050c4cbf3a1e7d0a378: Status 404 returned error can't find the container with id 40d87c7a01f3bd728ef41dee6e00a6fa7d102d7be1e78050c4cbf3a1e7d0a378 Oct 04 04:35:23 crc kubenswrapper[4770]: I1004 04:35:23.674356 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:35:23 crc kubenswrapper[4770]: E1004 04:35:23.675396 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:35:23 crc kubenswrapper[4770]: I1004 04:35:23.688193 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88594c99-e1c5-4f9f-b38b-fb125afbd5dc" path="/var/lib/kubelet/pods/88594c99-e1c5-4f9f-b38b-fb125afbd5dc/volumes" Oct 04 04:35:24 crc kubenswrapper[4770]: I1004 04:35:24.197872 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7783d14-4f52-47fa-8bb9-5a99cb5f7192","Type":"ContainerStarted","Data":"b2dd5c11848132ac3c875a3c06f8e1e90acea5f608bfb558073bd62ce045db50"} Oct 04 04:35:24 crc kubenswrapper[4770]: I1004 04:35:24.197909 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7783d14-4f52-47fa-8bb9-5a99cb5f7192","Type":"ContainerStarted","Data":"40d87c7a01f3bd728ef41dee6e00a6fa7d102d7be1e78050c4cbf3a1e7d0a378"} Oct 04 04:35:25 crc kubenswrapper[4770]: I1004 04:35:25.053354 4770 scope.go:117] "RemoveContainer" containerID="cefd1e82582080ec5b190383f292f3f0ad42766d4594318ecf8e72a163b9a2ac" Oct 04 04:35:25 crc kubenswrapper[4770]: I1004 04:35:25.210047 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7783d14-4f52-47fa-8bb9-5a99cb5f7192","Type":"ContainerStarted","Data":"5a5a976f6b1bcf77bbcbfdd404c05f4491480779ca55ecf4c5aa0037978849d7"} Oct 04 04:35:25 crc kubenswrapper[4770]: I1004 04:35:25.240467 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.240445324 podStartE2EDuration="3.240445324s" podCreationTimestamp="2025-10-04 04:35:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:25.234030836 +0000 UTC m=+5536.526040578" watchObservedRunningTime="2025-10-04 04:35:25.240445324 +0000 UTC m=+5536.532455036" Oct 04 04:35:25 crc kubenswrapper[4770]: I1004 04:35:25.864148 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:35:25 crc kubenswrapper[4770]: I1004 04:35:25.930784 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69f87f7665-xmlfg"] Oct 04 04:35:25 crc kubenswrapper[4770]: I1004 04:35:25.931085 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" podUID="b63b1f17-3ef0-4d65-8575-aea646f6c226" containerName="dnsmasq-dns" containerID="cri-o://672ef7ca17c64073cac265a3beb342e62dbe0e1c85567cba7a96a65a8343d071" gracePeriod=10 Oct 04 04:35:26 crc kubenswrapper[4770]: I1004 04:35:26.224993 4770 generic.go:334] "Generic (PLEG): container finished" podID="b63b1f17-3ef0-4d65-8575-aea646f6c226" containerID="672ef7ca17c64073cac265a3beb342e62dbe0e1c85567cba7a96a65a8343d071" exitCode=0 Oct 04 04:35:26 crc kubenswrapper[4770]: I1004 04:35:26.225102 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" event={"ID":"b63b1f17-3ef0-4d65-8575-aea646f6c226","Type":"ContainerDied","Data":"672ef7ca17c64073cac265a3beb342e62dbe0e1c85567cba7a96a65a8343d071"} Oct 04 04:35:26 crc kubenswrapper[4770]: I1004 04:35:26.941098 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.091525 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-dns-svc\") pod \"b63b1f17-3ef0-4d65-8575-aea646f6c226\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.091616 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6f4f\" (UniqueName: \"kubernetes.io/projected/b63b1f17-3ef0-4d65-8575-aea646f6c226-kube-api-access-k6f4f\") pod \"b63b1f17-3ef0-4d65-8575-aea646f6c226\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.091640 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-config\") pod \"b63b1f17-3ef0-4d65-8575-aea646f6c226\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.091729 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-nb\") pod \"b63b1f17-3ef0-4d65-8575-aea646f6c226\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.091797 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-sb\") pod \"b63b1f17-3ef0-4d65-8575-aea646f6c226\" (UID: \"b63b1f17-3ef0-4d65-8575-aea646f6c226\") " Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.099291 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b63b1f17-3ef0-4d65-8575-aea646f6c226-kube-api-access-k6f4f" (OuterVolumeSpecName: "kube-api-access-k6f4f") pod "b63b1f17-3ef0-4d65-8575-aea646f6c226" (UID: "b63b1f17-3ef0-4d65-8575-aea646f6c226"). InnerVolumeSpecName "kube-api-access-k6f4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.136688 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b63b1f17-3ef0-4d65-8575-aea646f6c226" (UID: "b63b1f17-3ef0-4d65-8575-aea646f6c226"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.138185 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-config" (OuterVolumeSpecName: "config") pod "b63b1f17-3ef0-4d65-8575-aea646f6c226" (UID: "b63b1f17-3ef0-4d65-8575-aea646f6c226"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.168276 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b63b1f17-3ef0-4d65-8575-aea646f6c226" (UID: "b63b1f17-3ef0-4d65-8575-aea646f6c226"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.172234 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b63b1f17-3ef0-4d65-8575-aea646f6c226" (UID: "b63b1f17-3ef0-4d65-8575-aea646f6c226"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.194249 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6f4f\" (UniqueName: \"kubernetes.io/projected/b63b1f17-3ef0-4d65-8575-aea646f6c226-kube-api-access-k6f4f\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.194290 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.194308 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.194326 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.194344 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b63b1f17-3ef0-4d65-8575-aea646f6c226-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.237935 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" event={"ID":"b63b1f17-3ef0-4d65-8575-aea646f6c226","Type":"ContainerDied","Data":"be1b4ad5f6c26f3d4e44cd5a7f1acd64c3a22f17087990264a012506f14c9897"} Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.237996 4770 scope.go:117] "RemoveContainer" containerID="672ef7ca17c64073cac265a3beb342e62dbe0e1c85567cba7a96a65a8343d071" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.239290 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.268220 4770 scope.go:117] "RemoveContainer" containerID="f199045482713d2a6a6c1cc671e5c1e7ccec3047d15c616b1417a07f34212049" Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.288731 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69f87f7665-xmlfg"] Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.296571 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69f87f7665-xmlfg"] Oct 04 04:35:27 crc kubenswrapper[4770]: I1004 04:35:27.688891 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b63b1f17-3ef0-4d65-8575-aea646f6c226" path="/var/lib/kubelet/pods/b63b1f17-3ef0-4d65-8575-aea646f6c226/volumes" Oct 04 04:35:29 crc kubenswrapper[4770]: I1004 04:35:29.515639 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:35:29 crc kubenswrapper[4770]: I1004 04:35:29.517845 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:35:29 crc kubenswrapper[4770]: I1004 04:35:29.571519 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:35:29 crc kubenswrapper[4770]: I1004 04:35:29.599181 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:35:30 crc kubenswrapper[4770]: I1004 04:35:30.276058 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:35:30 crc kubenswrapper[4770]: I1004 04:35:30.276117 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:35:31 crc kubenswrapper[4770]: I1004 04:35:31.892118 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-69f87f7665-xmlfg" podUID="b63b1f17-3ef0-4d65-8575-aea646f6c226" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.36:5353: i/o timeout" Oct 04 04:35:32 crc kubenswrapper[4770]: I1004 04:35:32.370704 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:35:32 crc kubenswrapper[4770]: I1004 04:35:32.370874 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:35:32 crc kubenswrapper[4770]: I1004 04:35:32.372715 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:35:32 crc kubenswrapper[4770]: I1004 04:35:32.561645 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:32 crc kubenswrapper[4770]: I1004 04:35:32.561722 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:32 crc kubenswrapper[4770]: I1004 04:35:32.593288 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:32 crc kubenswrapper[4770]: I1004 04:35:32.608915 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:33 crc kubenswrapper[4770]: I1004 04:35:33.306342 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:33 crc kubenswrapper[4770]: I1004 04:35:33.306857 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:34 crc kubenswrapper[4770]: I1004 04:35:34.673757 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:35:34 crc kubenswrapper[4770]: E1004 04:35:34.674266 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:35:35 crc kubenswrapper[4770]: I1004 04:35:35.363391 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:35 crc kubenswrapper[4770]: I1004 04:35:35.363891 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:35:35 crc kubenswrapper[4770]: I1004 04:35:35.365435 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:35:41 crc kubenswrapper[4770]: I1004 04:35:41.847690 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-dql4d"] Oct 04 04:35:41 crc kubenswrapper[4770]: E1004 04:35:41.849405 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b63b1f17-3ef0-4d65-8575-aea646f6c226" containerName="init" Oct 04 04:35:41 crc kubenswrapper[4770]: I1004 04:35:41.849483 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b63b1f17-3ef0-4d65-8575-aea646f6c226" containerName="init" Oct 04 04:35:41 crc kubenswrapper[4770]: E1004 04:35:41.849551 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b63b1f17-3ef0-4d65-8575-aea646f6c226" containerName="dnsmasq-dns" Oct 04 04:35:41 crc kubenswrapper[4770]: I1004 04:35:41.849608 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b63b1f17-3ef0-4d65-8575-aea646f6c226" containerName="dnsmasq-dns" Oct 04 04:35:41 crc kubenswrapper[4770]: I1004 04:35:41.849814 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b63b1f17-3ef0-4d65-8575-aea646f6c226" containerName="dnsmasq-dns" Oct 04 04:35:41 crc kubenswrapper[4770]: I1004 04:35:41.850445 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dql4d" Oct 04 04:35:41 crc kubenswrapper[4770]: I1004 04:35:41.863909 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-dql4d"] Oct 04 04:35:41 crc kubenswrapper[4770]: I1004 04:35:41.985110 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc7s4\" (UniqueName: \"kubernetes.io/projected/5c53d120-d264-4191-884d-e377d0ffbb2e-kube-api-access-rc7s4\") pod \"placement-db-create-dql4d\" (UID: \"5c53d120-d264-4191-884d-e377d0ffbb2e\") " pod="openstack/placement-db-create-dql4d" Oct 04 04:35:42 crc kubenswrapper[4770]: I1004 04:35:42.087075 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc7s4\" (UniqueName: \"kubernetes.io/projected/5c53d120-d264-4191-884d-e377d0ffbb2e-kube-api-access-rc7s4\") pod \"placement-db-create-dql4d\" (UID: \"5c53d120-d264-4191-884d-e377d0ffbb2e\") " pod="openstack/placement-db-create-dql4d" Oct 04 04:35:42 crc kubenswrapper[4770]: I1004 04:35:42.114826 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc7s4\" (UniqueName: \"kubernetes.io/projected/5c53d120-d264-4191-884d-e377d0ffbb2e-kube-api-access-rc7s4\") pod \"placement-db-create-dql4d\" (UID: \"5c53d120-d264-4191-884d-e377d0ffbb2e\") " pod="openstack/placement-db-create-dql4d" Oct 04 04:35:42 crc kubenswrapper[4770]: I1004 04:35:42.166887 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dql4d" Oct 04 04:35:42 crc kubenswrapper[4770]: I1004 04:35:42.621284 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-dql4d"] Oct 04 04:35:42 crc kubenswrapper[4770]: W1004 04:35:42.621932 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c53d120_d264_4191_884d_e377d0ffbb2e.slice/crio-10dffc2ee9e907efb455703f22ea9dfa6e7118a53a7412c8bb76638c0a51fdff WatchSource:0}: Error finding container 10dffc2ee9e907efb455703f22ea9dfa6e7118a53a7412c8bb76638c0a51fdff: Status 404 returned error can't find the container with id 10dffc2ee9e907efb455703f22ea9dfa6e7118a53a7412c8bb76638c0a51fdff Oct 04 04:35:43 crc kubenswrapper[4770]: I1004 04:35:43.394097 4770 generic.go:334] "Generic (PLEG): container finished" podID="5c53d120-d264-4191-884d-e377d0ffbb2e" containerID="fa5102de2ebaabe0371080b98993396ba4f3ba23ebe9e809050caafeceec55a3" exitCode=0 Oct 04 04:35:43 crc kubenswrapper[4770]: I1004 04:35:43.394274 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dql4d" event={"ID":"5c53d120-d264-4191-884d-e377d0ffbb2e","Type":"ContainerDied","Data":"fa5102de2ebaabe0371080b98993396ba4f3ba23ebe9e809050caafeceec55a3"} Oct 04 04:35:43 crc kubenswrapper[4770]: I1004 04:35:43.394350 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dql4d" event={"ID":"5c53d120-d264-4191-884d-e377d0ffbb2e","Type":"ContainerStarted","Data":"10dffc2ee9e907efb455703f22ea9dfa6e7118a53a7412c8bb76638c0a51fdff"} Oct 04 04:35:44 crc kubenswrapper[4770]: I1004 04:35:44.794901 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dql4d" Oct 04 04:35:44 crc kubenswrapper[4770]: I1004 04:35:44.954111 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc7s4\" (UniqueName: \"kubernetes.io/projected/5c53d120-d264-4191-884d-e377d0ffbb2e-kube-api-access-rc7s4\") pod \"5c53d120-d264-4191-884d-e377d0ffbb2e\" (UID: \"5c53d120-d264-4191-884d-e377d0ffbb2e\") " Oct 04 04:35:44 crc kubenswrapper[4770]: I1004 04:35:44.962398 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c53d120-d264-4191-884d-e377d0ffbb2e-kube-api-access-rc7s4" (OuterVolumeSpecName: "kube-api-access-rc7s4") pod "5c53d120-d264-4191-884d-e377d0ffbb2e" (UID: "5c53d120-d264-4191-884d-e377d0ffbb2e"). InnerVolumeSpecName "kube-api-access-rc7s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:45 crc kubenswrapper[4770]: I1004 04:35:45.055899 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc7s4\" (UniqueName: \"kubernetes.io/projected/5c53d120-d264-4191-884d-e377d0ffbb2e-kube-api-access-rc7s4\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:45 crc kubenswrapper[4770]: I1004 04:35:45.417837 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dql4d" event={"ID":"5c53d120-d264-4191-884d-e377d0ffbb2e","Type":"ContainerDied","Data":"10dffc2ee9e907efb455703f22ea9dfa6e7118a53a7412c8bb76638c0a51fdff"} Oct 04 04:35:45 crc kubenswrapper[4770]: I1004 04:35:45.417899 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10dffc2ee9e907efb455703f22ea9dfa6e7118a53a7412c8bb76638c0a51fdff" Oct 04 04:35:45 crc kubenswrapper[4770]: I1004 04:35:45.417973 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dql4d" Oct 04 04:35:49 crc kubenswrapper[4770]: I1004 04:35:49.687380 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:35:49 crc kubenswrapper[4770]: E1004 04:35:49.688599 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:35:51 crc kubenswrapper[4770]: I1004 04:35:51.877599 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7e7a-account-create-5nl84"] Oct 04 04:35:51 crc kubenswrapper[4770]: E1004 04:35:51.878402 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c53d120-d264-4191-884d-e377d0ffbb2e" containerName="mariadb-database-create" Oct 04 04:35:51 crc kubenswrapper[4770]: I1004 04:35:51.878419 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c53d120-d264-4191-884d-e377d0ffbb2e" containerName="mariadb-database-create" Oct 04 04:35:51 crc kubenswrapper[4770]: I1004 04:35:51.878646 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c53d120-d264-4191-884d-e377d0ffbb2e" containerName="mariadb-database-create" Oct 04 04:35:51 crc kubenswrapper[4770]: I1004 04:35:51.879389 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7e7a-account-create-5nl84" Oct 04 04:35:51 crc kubenswrapper[4770]: I1004 04:35:51.882483 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 04 04:35:51 crc kubenswrapper[4770]: I1004 04:35:51.906751 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7e7a-account-create-5nl84"] Oct 04 04:35:51 crc kubenswrapper[4770]: I1004 04:35:51.983978 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xf9q\" (UniqueName: \"kubernetes.io/projected/12c9dba9-f1c3-4402-8c94-39c3c36807f7-kube-api-access-5xf9q\") pod \"placement-7e7a-account-create-5nl84\" (UID: \"12c9dba9-f1c3-4402-8c94-39c3c36807f7\") " pod="openstack/placement-7e7a-account-create-5nl84" Oct 04 04:35:52 crc kubenswrapper[4770]: I1004 04:35:52.086325 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xf9q\" (UniqueName: \"kubernetes.io/projected/12c9dba9-f1c3-4402-8c94-39c3c36807f7-kube-api-access-5xf9q\") pod \"placement-7e7a-account-create-5nl84\" (UID: \"12c9dba9-f1c3-4402-8c94-39c3c36807f7\") " pod="openstack/placement-7e7a-account-create-5nl84" Oct 04 04:35:52 crc kubenswrapper[4770]: I1004 04:35:52.109457 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xf9q\" (UniqueName: \"kubernetes.io/projected/12c9dba9-f1c3-4402-8c94-39c3c36807f7-kube-api-access-5xf9q\") pod \"placement-7e7a-account-create-5nl84\" (UID: \"12c9dba9-f1c3-4402-8c94-39c3c36807f7\") " pod="openstack/placement-7e7a-account-create-5nl84" Oct 04 04:35:52 crc kubenswrapper[4770]: I1004 04:35:52.207420 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7e7a-account-create-5nl84" Oct 04 04:35:52 crc kubenswrapper[4770]: I1004 04:35:52.648273 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7e7a-account-create-5nl84"] Oct 04 04:35:53 crc kubenswrapper[4770]: I1004 04:35:53.511189 4770 generic.go:334] "Generic (PLEG): container finished" podID="12c9dba9-f1c3-4402-8c94-39c3c36807f7" containerID="a0a1678c9de94d202465457eaf1d439d1611f76c3acdb960b78874809b6cff85" exitCode=0 Oct 04 04:35:53 crc kubenswrapper[4770]: I1004 04:35:53.511273 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7e7a-account-create-5nl84" event={"ID":"12c9dba9-f1c3-4402-8c94-39c3c36807f7","Type":"ContainerDied","Data":"a0a1678c9de94d202465457eaf1d439d1611f76c3acdb960b78874809b6cff85"} Oct 04 04:35:53 crc kubenswrapper[4770]: I1004 04:35:53.511364 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7e7a-account-create-5nl84" event={"ID":"12c9dba9-f1c3-4402-8c94-39c3c36807f7","Type":"ContainerStarted","Data":"4e20fb6335dd541b0be9f35b128fa24f4ee649e2fe9dce3ea5a2c3bd7b2fcb1f"} Oct 04 04:35:54 crc kubenswrapper[4770]: I1004 04:35:54.888041 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7e7a-account-create-5nl84" Oct 04 04:35:55 crc kubenswrapper[4770]: I1004 04:35:55.039337 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xf9q\" (UniqueName: \"kubernetes.io/projected/12c9dba9-f1c3-4402-8c94-39c3c36807f7-kube-api-access-5xf9q\") pod \"12c9dba9-f1c3-4402-8c94-39c3c36807f7\" (UID: \"12c9dba9-f1c3-4402-8c94-39c3c36807f7\") " Oct 04 04:35:55 crc kubenswrapper[4770]: I1004 04:35:55.044929 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c9dba9-f1c3-4402-8c94-39c3c36807f7-kube-api-access-5xf9q" (OuterVolumeSpecName: "kube-api-access-5xf9q") pod "12c9dba9-f1c3-4402-8c94-39c3c36807f7" (UID: "12c9dba9-f1c3-4402-8c94-39c3c36807f7"). InnerVolumeSpecName "kube-api-access-5xf9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:55 crc kubenswrapper[4770]: I1004 04:35:55.143361 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xf9q\" (UniqueName: \"kubernetes.io/projected/12c9dba9-f1c3-4402-8c94-39c3c36807f7-kube-api-access-5xf9q\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:55 crc kubenswrapper[4770]: I1004 04:35:55.534096 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7e7a-account-create-5nl84" event={"ID":"12c9dba9-f1c3-4402-8c94-39c3c36807f7","Type":"ContainerDied","Data":"4e20fb6335dd541b0be9f35b128fa24f4ee649e2fe9dce3ea5a2c3bd7b2fcb1f"} Oct 04 04:35:55 crc kubenswrapper[4770]: I1004 04:35:55.534138 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e20fb6335dd541b0be9f35b128fa24f4ee649e2fe9dce3ea5a2c3bd7b2fcb1f" Oct 04 04:35:55 crc kubenswrapper[4770]: I1004 04:35:55.534174 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7e7a-account-create-5nl84" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.183302 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85467b57bf-hjkz7"] Oct 04 04:35:57 crc kubenswrapper[4770]: E1004 04:35:57.183996 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c9dba9-f1c3-4402-8c94-39c3c36807f7" containerName="mariadb-account-create" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.184042 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c9dba9-f1c3-4402-8c94-39c3c36807f7" containerName="mariadb-account-create" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.184261 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c9dba9-f1c3-4402-8c94-39c3c36807f7" containerName="mariadb-account-create" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.185329 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.214977 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85467b57bf-hjkz7"] Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.254834 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-cmdwr"] Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.256327 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.259819 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.259838 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9kl5v" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.260191 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.262244 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cmdwr"] Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.294261 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwjgp\" (UniqueName: \"kubernetes.io/projected/455e4130-03c6-4e06-a9f5-4df56558327f-kube-api-access-nwjgp\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.298217 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-dns-svc\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.300499 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-sb\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.300720 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-config\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.300845 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-nb\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.403788 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-config\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.403844 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-nb\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.403902 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rflfm\" (UniqueName: \"kubernetes.io/projected/aab0c547-9737-4e63-a6d4-fba02c488986-kube-api-access-rflfm\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.403924 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-config-data\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.403955 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab0c547-9737-4e63-a6d4-fba02c488986-logs\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.403978 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwjgp\" (UniqueName: \"kubernetes.io/projected/455e4130-03c6-4e06-a9f5-4df56558327f-kube-api-access-nwjgp\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.404091 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-combined-ca-bundle\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.404116 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-dns-svc\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.404143 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-sb\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.404182 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-scripts\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.405119 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-config\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.405207 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-dns-svc\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.405541 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-sb\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.405577 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-nb\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.420916 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwjgp\" (UniqueName: \"kubernetes.io/projected/455e4130-03c6-4e06-a9f5-4df56558327f-kube-api-access-nwjgp\") pod \"dnsmasq-dns-85467b57bf-hjkz7\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.506186 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-combined-ca-bundle\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.506792 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-scripts\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.506924 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rflfm\" (UniqueName: \"kubernetes.io/projected/aab0c547-9737-4e63-a6d4-fba02c488986-kube-api-access-rflfm\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.507052 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-config-data\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.507484 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab0c547-9737-4e63-a6d4-fba02c488986-logs\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.508120 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab0c547-9737-4e63-a6d4-fba02c488986-logs\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.511050 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-scripts\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.511333 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-config-data\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.511809 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-combined-ca-bundle\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.514843 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.527370 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rflfm\" (UniqueName: \"kubernetes.io/projected/aab0c547-9737-4e63-a6d4-fba02c488986-kube-api-access-rflfm\") pod \"placement-db-sync-cmdwr\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:57 crc kubenswrapper[4770]: I1004 04:35:57.581843 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cmdwr" Oct 04 04:35:58 crc kubenswrapper[4770]: I1004 04:35:58.315517 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85467b57bf-hjkz7"] Oct 04 04:35:58 crc kubenswrapper[4770]: I1004 04:35:58.375307 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cmdwr"] Oct 04 04:35:58 crc kubenswrapper[4770]: I1004 04:35:58.558266 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" event={"ID":"455e4130-03c6-4e06-a9f5-4df56558327f","Type":"ContainerStarted","Data":"b9f2a9209d86c861b178c1e6e31e60cc2c19d7b6221eba5ef9d87029d0af9eea"} Oct 04 04:35:58 crc kubenswrapper[4770]: I1004 04:35:58.559698 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cmdwr" event={"ID":"aab0c547-9737-4e63-a6d4-fba02c488986","Type":"ContainerStarted","Data":"3fad429f72fb578b567891f71d3aadb354ed8eace885ae35739174fefe867b2d"} Oct 04 04:35:59 crc kubenswrapper[4770]: I1004 04:35:59.576703 4770 generic.go:334] "Generic (PLEG): container finished" podID="455e4130-03c6-4e06-a9f5-4df56558327f" containerID="606e187a5ce7dba44d4054c6ecc4c8597e9c2eadeedde24e024b4cbb0fe7fd7a" exitCode=0 Oct 04 04:35:59 crc kubenswrapper[4770]: I1004 04:35:59.576769 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" event={"ID":"455e4130-03c6-4e06-a9f5-4df56558327f","Type":"ContainerDied","Data":"606e187a5ce7dba44d4054c6ecc4c8597e9c2eadeedde24e024b4cbb0fe7fd7a"} Oct 04 04:35:59 crc kubenswrapper[4770]: I1004 04:35:59.579127 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cmdwr" event={"ID":"aab0c547-9737-4e63-a6d4-fba02c488986","Type":"ContainerStarted","Data":"42dab1f26648852abc8e6b4050123a7f8b38fc0729b4bdfeef04dfa7472811c3"} Oct 04 04:35:59 crc kubenswrapper[4770]: I1004 04:35:59.641055 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-cmdwr" podStartSLOduration=2.64099116 podStartE2EDuration="2.64099116s" podCreationTimestamp="2025-10-04 04:35:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:59.622718002 +0000 UTC m=+5570.914727724" watchObservedRunningTime="2025-10-04 04:35:59.64099116 +0000 UTC m=+5570.933000902" Oct 04 04:36:00 crc kubenswrapper[4770]: I1004 04:36:00.592344 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" event={"ID":"455e4130-03c6-4e06-a9f5-4df56558327f","Type":"ContainerStarted","Data":"3b435a70ab461749e918093686c1a3f680f394bb87307f2df6c03621b997870d"} Oct 04 04:36:00 crc kubenswrapper[4770]: I1004 04:36:00.613600 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" podStartSLOduration=3.613578528 podStartE2EDuration="3.613578528s" podCreationTimestamp="2025-10-04 04:35:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:00.60827349 +0000 UTC m=+5571.900283202" watchObservedRunningTime="2025-10-04 04:36:00.613578528 +0000 UTC m=+5571.905588240" Oct 04 04:36:01 crc kubenswrapper[4770]: I1004 04:36:01.602101 4770 generic.go:334] "Generic (PLEG): container finished" podID="aab0c547-9737-4e63-a6d4-fba02c488986" containerID="42dab1f26648852abc8e6b4050123a7f8b38fc0729b4bdfeef04dfa7472811c3" exitCode=0 Oct 04 04:36:01 crc kubenswrapper[4770]: I1004 04:36:01.602188 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cmdwr" event={"ID":"aab0c547-9737-4e63-a6d4-fba02c488986","Type":"ContainerDied","Data":"42dab1f26648852abc8e6b4050123a7f8b38fc0729b4bdfeef04dfa7472811c3"} Oct 04 04:36:01 crc kubenswrapper[4770]: I1004 04:36:01.602510 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:36:01 crc kubenswrapper[4770]: I1004 04:36:01.673524 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:36:01 crc kubenswrapper[4770]: E1004 04:36:01.674259 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:36:02 crc kubenswrapper[4770]: I1004 04:36:02.944100 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cmdwr" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.009132 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-scripts\") pod \"aab0c547-9737-4e63-a6d4-fba02c488986\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.009205 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-combined-ca-bundle\") pod \"aab0c547-9737-4e63-a6d4-fba02c488986\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.009270 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-config-data\") pod \"aab0c547-9737-4e63-a6d4-fba02c488986\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.009321 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rflfm\" (UniqueName: \"kubernetes.io/projected/aab0c547-9737-4e63-a6d4-fba02c488986-kube-api-access-rflfm\") pod \"aab0c547-9737-4e63-a6d4-fba02c488986\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.009355 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab0c547-9737-4e63-a6d4-fba02c488986-logs\") pod \"aab0c547-9737-4e63-a6d4-fba02c488986\" (UID: \"aab0c547-9737-4e63-a6d4-fba02c488986\") " Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.009799 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aab0c547-9737-4e63-a6d4-fba02c488986-logs" (OuterVolumeSpecName: "logs") pod "aab0c547-9737-4e63-a6d4-fba02c488986" (UID: "aab0c547-9737-4e63-a6d4-fba02c488986"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.014250 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-scripts" (OuterVolumeSpecName: "scripts") pod "aab0c547-9737-4e63-a6d4-fba02c488986" (UID: "aab0c547-9737-4e63-a6d4-fba02c488986"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.014726 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aab0c547-9737-4e63-a6d4-fba02c488986-kube-api-access-rflfm" (OuterVolumeSpecName: "kube-api-access-rflfm") pod "aab0c547-9737-4e63-a6d4-fba02c488986" (UID: "aab0c547-9737-4e63-a6d4-fba02c488986"). InnerVolumeSpecName "kube-api-access-rflfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.032970 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-config-data" (OuterVolumeSpecName: "config-data") pod "aab0c547-9737-4e63-a6d4-fba02c488986" (UID: "aab0c547-9737-4e63-a6d4-fba02c488986"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.034263 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aab0c547-9737-4e63-a6d4-fba02c488986" (UID: "aab0c547-9737-4e63-a6d4-fba02c488986"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.112580 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.112615 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.112631 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab0c547-9737-4e63-a6d4-fba02c488986-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.112645 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rflfm\" (UniqueName: \"kubernetes.io/projected/aab0c547-9737-4e63-a6d4-fba02c488986-kube-api-access-rflfm\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.112657 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aab0c547-9737-4e63-a6d4-fba02c488986-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.622152 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cmdwr" event={"ID":"aab0c547-9737-4e63-a6d4-fba02c488986","Type":"ContainerDied","Data":"3fad429f72fb578b567891f71d3aadb354ed8eace885ae35739174fefe867b2d"} Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.622199 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fad429f72fb578b567891f71d3aadb354ed8eace885ae35739174fefe867b2d" Oct 04 04:36:03 crc kubenswrapper[4770]: I1004 04:36:03.622268 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cmdwr" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.132550 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-b475498cb-2njbg"] Oct 04 04:36:04 crc kubenswrapper[4770]: E1004 04:36:04.133297 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab0c547-9737-4e63-a6d4-fba02c488986" containerName="placement-db-sync" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.133316 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab0c547-9737-4e63-a6d4-fba02c488986" containerName="placement-db-sync" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.133563 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab0c547-9737-4e63-a6d4-fba02c488986" containerName="placement-db-sync" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.135112 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.138164 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.138425 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.140678 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9kl5v" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.146085 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b475498cb-2njbg"] Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.335960 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ced10f27-425f-4a56-9d11-9bce2a12a8ac-logs\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.336075 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcghf\" (UniqueName: \"kubernetes.io/projected/ced10f27-425f-4a56-9d11-9bce2a12a8ac-kube-api-access-pcghf\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.336128 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced10f27-425f-4a56-9d11-9bce2a12a8ac-combined-ca-bundle\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.336203 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced10f27-425f-4a56-9d11-9bce2a12a8ac-config-data\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.336254 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced10f27-425f-4a56-9d11-9bce2a12a8ac-scripts\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.437268 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ced10f27-425f-4a56-9d11-9bce2a12a8ac-logs\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.437342 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcghf\" (UniqueName: \"kubernetes.io/projected/ced10f27-425f-4a56-9d11-9bce2a12a8ac-kube-api-access-pcghf\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.437364 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced10f27-425f-4a56-9d11-9bce2a12a8ac-combined-ca-bundle\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.437418 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced10f27-425f-4a56-9d11-9bce2a12a8ac-config-data\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.437452 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced10f27-425f-4a56-9d11-9bce2a12a8ac-scripts\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.438158 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ced10f27-425f-4a56-9d11-9bce2a12a8ac-logs\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.442410 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ced10f27-425f-4a56-9d11-9bce2a12a8ac-combined-ca-bundle\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.446779 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ced10f27-425f-4a56-9d11-9bce2a12a8ac-config-data\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.451328 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ced10f27-425f-4a56-9d11-9bce2a12a8ac-scripts\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.456960 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcghf\" (UniqueName: \"kubernetes.io/projected/ced10f27-425f-4a56-9d11-9bce2a12a8ac-kube-api-access-pcghf\") pod \"placement-b475498cb-2njbg\" (UID: \"ced10f27-425f-4a56-9d11-9bce2a12a8ac\") " pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:04 crc kubenswrapper[4770]: I1004 04:36:04.753368 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:05 crc kubenswrapper[4770]: I1004 04:36:05.228753 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b475498cb-2njbg"] Oct 04 04:36:05 crc kubenswrapper[4770]: I1004 04:36:05.642092 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b475498cb-2njbg" event={"ID":"ced10f27-425f-4a56-9d11-9bce2a12a8ac","Type":"ContainerStarted","Data":"e6bf8d1dede1c2f45365c899aa4964652c84fe5f7582f171ea083724467b1617"} Oct 04 04:36:05 crc kubenswrapper[4770]: I1004 04:36:05.642428 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b475498cb-2njbg" event={"ID":"ced10f27-425f-4a56-9d11-9bce2a12a8ac","Type":"ContainerStarted","Data":"5b7778225955b071ad67ec497acf00ddd9d2b515e540ca33a5524b8b58b7d140"} Oct 04 04:36:06 crc kubenswrapper[4770]: I1004 04:36:06.656454 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b475498cb-2njbg" event={"ID":"ced10f27-425f-4a56-9d11-9bce2a12a8ac","Type":"ContainerStarted","Data":"4ec791418e5476fc2ee0b344c2416f44340e1334d45f831818d11395cb0a5c39"} Oct 04 04:36:06 crc kubenswrapper[4770]: I1004 04:36:06.656991 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:06 crc kubenswrapper[4770]: I1004 04:36:06.657035 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:06 crc kubenswrapper[4770]: I1004 04:36:06.681548 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-b475498cb-2njbg" podStartSLOduration=2.68152256 podStartE2EDuration="2.68152256s" podCreationTimestamp="2025-10-04 04:36:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:36:06.672781192 +0000 UTC m=+5577.964790914" watchObservedRunningTime="2025-10-04 04:36:06.68152256 +0000 UTC m=+5577.973532292" Oct 04 04:36:07 crc kubenswrapper[4770]: I1004 04:36:07.516244 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:36:07 crc kubenswrapper[4770]: I1004 04:36:07.575483 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d4d6fcd69-chhqb"] Oct 04 04:36:07 crc kubenswrapper[4770]: I1004 04:36:07.575710 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" podUID="0684e72c-e2a5-4982-8f16-8cc8b10a152a" containerName="dnsmasq-dns" containerID="cri-o://8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d" gracePeriod=10 Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.113236 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.300828 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-nb\") pod \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.300920 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-dns-svc\") pod \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.300966 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-config\") pod \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.301140 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-sb\") pod \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.301189 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6bf9\" (UniqueName: \"kubernetes.io/projected/0684e72c-e2a5-4982-8f16-8cc8b10a152a-kube-api-access-v6bf9\") pod \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\" (UID: \"0684e72c-e2a5-4982-8f16-8cc8b10a152a\") " Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.312248 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0684e72c-e2a5-4982-8f16-8cc8b10a152a-kube-api-access-v6bf9" (OuterVolumeSpecName: "kube-api-access-v6bf9") pod "0684e72c-e2a5-4982-8f16-8cc8b10a152a" (UID: "0684e72c-e2a5-4982-8f16-8cc8b10a152a"). InnerVolumeSpecName "kube-api-access-v6bf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.355195 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0684e72c-e2a5-4982-8f16-8cc8b10a152a" (UID: "0684e72c-e2a5-4982-8f16-8cc8b10a152a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.359621 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0684e72c-e2a5-4982-8f16-8cc8b10a152a" (UID: "0684e72c-e2a5-4982-8f16-8cc8b10a152a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.370695 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0684e72c-e2a5-4982-8f16-8cc8b10a152a" (UID: "0684e72c-e2a5-4982-8f16-8cc8b10a152a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.371777 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-config" (OuterVolumeSpecName: "config") pod "0684e72c-e2a5-4982-8f16-8cc8b10a152a" (UID: "0684e72c-e2a5-4982-8f16-8cc8b10a152a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.402653 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.402701 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6bf9\" (UniqueName: \"kubernetes.io/projected/0684e72c-e2a5-4982-8f16-8cc8b10a152a-kube-api-access-v6bf9\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.402715 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.402728 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.402740 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0684e72c-e2a5-4982-8f16-8cc8b10a152a-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.677450 4770 generic.go:334] "Generic (PLEG): container finished" podID="0684e72c-e2a5-4982-8f16-8cc8b10a152a" containerID="8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d" exitCode=0 Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.677510 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" event={"ID":"0684e72c-e2a5-4982-8f16-8cc8b10a152a","Type":"ContainerDied","Data":"8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d"} Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.677580 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" event={"ID":"0684e72c-e2a5-4982-8f16-8cc8b10a152a","Type":"ContainerDied","Data":"5b805fbd6c67e3cff5f09c27975418e2b8c9758836666df7b6412a4c68435d7f"} Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.677588 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4d6fcd69-chhqb" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.677607 4770 scope.go:117] "RemoveContainer" containerID="8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.710535 4770 scope.go:117] "RemoveContainer" containerID="a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.720101 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d4d6fcd69-chhqb"] Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.726192 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d4d6fcd69-chhqb"] Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.738513 4770 scope.go:117] "RemoveContainer" containerID="8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d" Oct 04 04:36:08 crc kubenswrapper[4770]: E1004 04:36:08.738890 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d\": container with ID starting with 8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d not found: ID does not exist" containerID="8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.738935 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d"} err="failed to get container status \"8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d\": rpc error: code = NotFound desc = could not find container \"8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d\": container with ID starting with 8969d21ae090af4eab56424bb8ffa78926e8ca47dde3b4cba9d8668f45b3db7d not found: ID does not exist" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.738961 4770 scope.go:117] "RemoveContainer" containerID="a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad" Oct 04 04:36:08 crc kubenswrapper[4770]: E1004 04:36:08.739403 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad\": container with ID starting with a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad not found: ID does not exist" containerID="a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad" Oct 04 04:36:08 crc kubenswrapper[4770]: I1004 04:36:08.739439 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad"} err="failed to get container status \"a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad\": rpc error: code = NotFound desc = could not find container \"a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad\": container with ID starting with a470b96fa5b318f12ae4c8c5cc1cea82f0ab1024ca6319d2f24b1f4e9f6811ad not found: ID does not exist" Oct 04 04:36:09 crc kubenswrapper[4770]: I1004 04:36:09.703387 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0684e72c-e2a5-4982-8f16-8cc8b10a152a" path="/var/lib/kubelet/pods/0684e72c-e2a5-4982-8f16-8cc8b10a152a/volumes" Oct 04 04:36:16 crc kubenswrapper[4770]: I1004 04:36:16.673349 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:36:16 crc kubenswrapper[4770]: E1004 04:36:16.674132 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:36:25 crc kubenswrapper[4770]: I1004 04:36:25.156822 4770 scope.go:117] "RemoveContainer" containerID="f7ade7e5774951aa70178dbd6a25a6da0f5dbef79302330caae5301b1234ae2f" Oct 04 04:36:25 crc kubenswrapper[4770]: I1004 04:36:25.187320 4770 scope.go:117] "RemoveContainer" containerID="6cb4056e8832a1959563883de74266a80f382bfddd48a1b55ccb6a0ea3a8481d" Oct 04 04:36:25 crc kubenswrapper[4770]: I1004 04:36:25.246605 4770 scope.go:117] "RemoveContainer" containerID="4efa024db728ef84763468355c2e43f41791002f10d9b2e2b1363c5d5a033ea0" Oct 04 04:36:30 crc kubenswrapper[4770]: I1004 04:36:30.674337 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:36:30 crc kubenswrapper[4770]: E1004 04:36:30.675078 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:36:36 crc kubenswrapper[4770]: I1004 04:36:36.088416 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:36 crc kubenswrapper[4770]: I1004 04:36:36.158143 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-b475498cb-2njbg" Oct 04 04:36:45 crc kubenswrapper[4770]: I1004 04:36:45.673579 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:36:46 crc kubenswrapper[4770]: I1004 04:36:46.057881 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"2e30f9d379ccb3500c52f46d443a65b8cc328fea98419caf04a4e9a30d4bd0cb"} Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.286262 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-js2xf"] Oct 04 04:36:59 crc kubenswrapper[4770]: E1004 04:36:59.287170 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0684e72c-e2a5-4982-8f16-8cc8b10a152a" containerName="dnsmasq-dns" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.287184 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0684e72c-e2a5-4982-8f16-8cc8b10a152a" containerName="dnsmasq-dns" Oct 04 04:36:59 crc kubenswrapper[4770]: E1004 04:36:59.287196 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0684e72c-e2a5-4982-8f16-8cc8b10a152a" containerName="init" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.287201 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0684e72c-e2a5-4982-8f16-8cc8b10a152a" containerName="init" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.287368 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0684e72c-e2a5-4982-8f16-8cc8b10a152a" containerName="dnsmasq-dns" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.287950 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-js2xf" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.298568 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-js2xf"] Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.342826 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz8wb\" (UniqueName: \"kubernetes.io/projected/2c3bbbec-3d17-487c-b37d-4ed7254d14f5-kube-api-access-zz8wb\") pod \"nova-api-db-create-js2xf\" (UID: \"2c3bbbec-3d17-487c-b37d-4ed7254d14f5\") " pod="openstack/nova-api-db-create-js2xf" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.373725 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-cwx5p"] Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.375239 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cwx5p" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.382159 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cwx5p"] Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.444127 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz8wb\" (UniqueName: \"kubernetes.io/projected/2c3bbbec-3d17-487c-b37d-4ed7254d14f5-kube-api-access-zz8wb\") pod \"nova-api-db-create-js2xf\" (UID: \"2c3bbbec-3d17-487c-b37d-4ed7254d14f5\") " pod="openstack/nova-api-db-create-js2xf" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.444190 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-784tz\" (UniqueName: \"kubernetes.io/projected/322194fd-446b-49b0-9051-20edd7db6d8a-kube-api-access-784tz\") pod \"nova-cell0-db-create-cwx5p\" (UID: \"322194fd-446b-49b0-9051-20edd7db6d8a\") " pod="openstack/nova-cell0-db-create-cwx5p" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.492701 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz8wb\" (UniqueName: \"kubernetes.io/projected/2c3bbbec-3d17-487c-b37d-4ed7254d14f5-kube-api-access-zz8wb\") pod \"nova-api-db-create-js2xf\" (UID: \"2c3bbbec-3d17-487c-b37d-4ed7254d14f5\") " pod="openstack/nova-api-db-create-js2xf" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.509576 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-jpf9d"] Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.516660 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jpf9d" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.547152 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-784tz\" (UniqueName: \"kubernetes.io/projected/322194fd-446b-49b0-9051-20edd7db6d8a-kube-api-access-784tz\") pod \"nova-cell0-db-create-cwx5p\" (UID: \"322194fd-446b-49b0-9051-20edd7db6d8a\") " pod="openstack/nova-cell0-db-create-cwx5p" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.574474 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jpf9d"] Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.578783 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-784tz\" (UniqueName: \"kubernetes.io/projected/322194fd-446b-49b0-9051-20edd7db6d8a-kube-api-access-784tz\") pod \"nova-cell0-db-create-cwx5p\" (UID: \"322194fd-446b-49b0-9051-20edd7db6d8a\") " pod="openstack/nova-cell0-db-create-cwx5p" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.617076 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-js2xf" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.649181 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlvqd\" (UniqueName: \"kubernetes.io/projected/2e9b805f-4925-4291-9a15-a7efac711ba5-kube-api-access-qlvqd\") pod \"nova-cell1-db-create-jpf9d\" (UID: \"2e9b805f-4925-4291-9a15-a7efac711ba5\") " pod="openstack/nova-cell1-db-create-jpf9d" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.690709 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cwx5p" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.751021 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlvqd\" (UniqueName: \"kubernetes.io/projected/2e9b805f-4925-4291-9a15-a7efac711ba5-kube-api-access-qlvqd\") pod \"nova-cell1-db-create-jpf9d\" (UID: \"2e9b805f-4925-4291-9a15-a7efac711ba5\") " pod="openstack/nova-cell1-db-create-jpf9d" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.770918 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlvqd\" (UniqueName: \"kubernetes.io/projected/2e9b805f-4925-4291-9a15-a7efac711ba5-kube-api-access-qlvqd\") pod \"nova-cell1-db-create-jpf9d\" (UID: \"2e9b805f-4925-4291-9a15-a7efac711ba5\") " pod="openstack/nova-cell1-db-create-jpf9d" Oct 04 04:36:59 crc kubenswrapper[4770]: I1004 04:36:59.849323 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jpf9d" Oct 04 04:37:00 crc kubenswrapper[4770]: I1004 04:37:00.043049 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cwx5p"] Oct 04 04:37:00 crc kubenswrapper[4770]: I1004 04:37:00.149577 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-js2xf"] Oct 04 04:37:00 crc kubenswrapper[4770]: W1004 04:37:00.153569 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c3bbbec_3d17_487c_b37d_4ed7254d14f5.slice/crio-613a8b5be18a31b3584486941632593a9873467269433e9733b038bf4de2ca4e WatchSource:0}: Error finding container 613a8b5be18a31b3584486941632593a9873467269433e9733b038bf4de2ca4e: Status 404 returned error can't find the container with id 613a8b5be18a31b3584486941632593a9873467269433e9733b038bf4de2ca4e Oct 04 04:37:00 crc kubenswrapper[4770]: I1004 04:37:00.201341 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-js2xf" event={"ID":"2c3bbbec-3d17-487c-b37d-4ed7254d14f5","Type":"ContainerStarted","Data":"613a8b5be18a31b3584486941632593a9873467269433e9733b038bf4de2ca4e"} Oct 04 04:37:00 crc kubenswrapper[4770]: I1004 04:37:00.203550 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cwx5p" event={"ID":"322194fd-446b-49b0-9051-20edd7db6d8a","Type":"ContainerStarted","Data":"3455bda5d1f89b2083c36030413f632756df26d726e21bedd7bb9978473bfb7f"} Oct 04 04:37:00 crc kubenswrapper[4770]: I1004 04:37:00.346348 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jpf9d"] Oct 04 04:37:00 crc kubenswrapper[4770]: W1004 04:37:00.349706 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e9b805f_4925_4291_9a15_a7efac711ba5.slice/crio-1262cbfd8830e3d72129e9e33f64f5c526515446c6bb49e30e3687474d35a718 WatchSource:0}: Error finding container 1262cbfd8830e3d72129e9e33f64f5c526515446c6bb49e30e3687474d35a718: Status 404 returned error can't find the container with id 1262cbfd8830e3d72129e9e33f64f5c526515446c6bb49e30e3687474d35a718 Oct 04 04:37:01 crc kubenswrapper[4770]: I1004 04:37:01.212600 4770 generic.go:334] "Generic (PLEG): container finished" podID="2e9b805f-4925-4291-9a15-a7efac711ba5" containerID="ca76c67b2fc3d2abd623444b7f7d0deefd3a2f76d63a806f53f7693a0672e3b7" exitCode=0 Oct 04 04:37:01 crc kubenswrapper[4770]: I1004 04:37:01.212665 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jpf9d" event={"ID":"2e9b805f-4925-4291-9a15-a7efac711ba5","Type":"ContainerDied","Data":"ca76c67b2fc3d2abd623444b7f7d0deefd3a2f76d63a806f53f7693a0672e3b7"} Oct 04 04:37:01 crc kubenswrapper[4770]: I1004 04:37:01.212955 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jpf9d" event={"ID":"2e9b805f-4925-4291-9a15-a7efac711ba5","Type":"ContainerStarted","Data":"1262cbfd8830e3d72129e9e33f64f5c526515446c6bb49e30e3687474d35a718"} Oct 04 04:37:01 crc kubenswrapper[4770]: I1004 04:37:01.214978 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-js2xf" event={"ID":"2c3bbbec-3d17-487c-b37d-4ed7254d14f5","Type":"ContainerStarted","Data":"8c51acb01d6a9084308d50feea9fe0f3d2cafa976899da2f19aabac10056472a"} Oct 04 04:37:01 crc kubenswrapper[4770]: I1004 04:37:01.216404 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cwx5p" event={"ID":"322194fd-446b-49b0-9051-20edd7db6d8a","Type":"ContainerStarted","Data":"a1ab32e4e547c7eed4f93a99d3690074701c89969f02b2c1201937c5a14cf3be"} Oct 04 04:37:01 crc kubenswrapper[4770]: I1004 04:37:01.244577 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-cwx5p" podStartSLOduration=2.244561456 podStartE2EDuration="2.244561456s" podCreationTimestamp="2025-10-04 04:36:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:01.241935607 +0000 UTC m=+5632.533945329" watchObservedRunningTime="2025-10-04 04:37:01.244561456 +0000 UTC m=+5632.536571168" Oct 04 04:37:01 crc kubenswrapper[4770]: I1004 04:37:01.268283 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-js2xf" podStartSLOduration=2.268261265 podStartE2EDuration="2.268261265s" podCreationTimestamp="2025-10-04 04:36:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:01.258031807 +0000 UTC m=+5632.550041519" watchObservedRunningTime="2025-10-04 04:37:01.268261265 +0000 UTC m=+5632.560270977" Oct 04 04:37:02 crc kubenswrapper[4770]: I1004 04:37:02.229614 4770 generic.go:334] "Generic (PLEG): container finished" podID="2c3bbbec-3d17-487c-b37d-4ed7254d14f5" containerID="8c51acb01d6a9084308d50feea9fe0f3d2cafa976899da2f19aabac10056472a" exitCode=0 Oct 04 04:37:02 crc kubenswrapper[4770]: I1004 04:37:02.229728 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-js2xf" event={"ID":"2c3bbbec-3d17-487c-b37d-4ed7254d14f5","Type":"ContainerDied","Data":"8c51acb01d6a9084308d50feea9fe0f3d2cafa976899da2f19aabac10056472a"} Oct 04 04:37:02 crc kubenswrapper[4770]: I1004 04:37:02.232320 4770 generic.go:334] "Generic (PLEG): container finished" podID="322194fd-446b-49b0-9051-20edd7db6d8a" containerID="a1ab32e4e547c7eed4f93a99d3690074701c89969f02b2c1201937c5a14cf3be" exitCode=0 Oct 04 04:37:02 crc kubenswrapper[4770]: I1004 04:37:02.232612 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cwx5p" event={"ID":"322194fd-446b-49b0-9051-20edd7db6d8a","Type":"ContainerDied","Data":"a1ab32e4e547c7eed4f93a99d3690074701c89969f02b2c1201937c5a14cf3be"} Oct 04 04:37:02 crc kubenswrapper[4770]: I1004 04:37:02.593855 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jpf9d" Oct 04 04:37:02 crc kubenswrapper[4770]: I1004 04:37:02.710481 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlvqd\" (UniqueName: \"kubernetes.io/projected/2e9b805f-4925-4291-9a15-a7efac711ba5-kube-api-access-qlvqd\") pod \"2e9b805f-4925-4291-9a15-a7efac711ba5\" (UID: \"2e9b805f-4925-4291-9a15-a7efac711ba5\") " Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.040380 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e9b805f-4925-4291-9a15-a7efac711ba5-kube-api-access-qlvqd" (OuterVolumeSpecName: "kube-api-access-qlvqd") pod "2e9b805f-4925-4291-9a15-a7efac711ba5" (UID: "2e9b805f-4925-4291-9a15-a7efac711ba5"). InnerVolumeSpecName "kube-api-access-qlvqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.118570 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlvqd\" (UniqueName: \"kubernetes.io/projected/2e9b805f-4925-4291-9a15-a7efac711ba5-kube-api-access-qlvqd\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.243565 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jpf9d" Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.245394 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jpf9d" event={"ID":"2e9b805f-4925-4291-9a15-a7efac711ba5","Type":"ContainerDied","Data":"1262cbfd8830e3d72129e9e33f64f5c526515446c6bb49e30e3687474d35a718"} Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.245451 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1262cbfd8830e3d72129e9e33f64f5c526515446c6bb49e30e3687474d35a718" Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.654105 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-js2xf" Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.734234 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz8wb\" (UniqueName: \"kubernetes.io/projected/2c3bbbec-3d17-487c-b37d-4ed7254d14f5-kube-api-access-zz8wb\") pod \"2c3bbbec-3d17-487c-b37d-4ed7254d14f5\" (UID: \"2c3bbbec-3d17-487c-b37d-4ed7254d14f5\") " Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.747237 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c3bbbec-3d17-487c-b37d-4ed7254d14f5-kube-api-access-zz8wb" (OuterVolumeSpecName: "kube-api-access-zz8wb") pod "2c3bbbec-3d17-487c-b37d-4ed7254d14f5" (UID: "2c3bbbec-3d17-487c-b37d-4ed7254d14f5"). InnerVolumeSpecName "kube-api-access-zz8wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.837072 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz8wb\" (UniqueName: \"kubernetes.io/projected/2c3bbbec-3d17-487c-b37d-4ed7254d14f5-kube-api-access-zz8wb\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:03 crc kubenswrapper[4770]: I1004 04:37:03.944338 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cwx5p" Oct 04 04:37:04 crc kubenswrapper[4770]: I1004 04:37:04.040241 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-784tz\" (UniqueName: \"kubernetes.io/projected/322194fd-446b-49b0-9051-20edd7db6d8a-kube-api-access-784tz\") pod \"322194fd-446b-49b0-9051-20edd7db6d8a\" (UID: \"322194fd-446b-49b0-9051-20edd7db6d8a\") " Oct 04 04:37:04 crc kubenswrapper[4770]: I1004 04:37:04.046233 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/322194fd-446b-49b0-9051-20edd7db6d8a-kube-api-access-784tz" (OuterVolumeSpecName: "kube-api-access-784tz") pod "322194fd-446b-49b0-9051-20edd7db6d8a" (UID: "322194fd-446b-49b0-9051-20edd7db6d8a"). InnerVolumeSpecName "kube-api-access-784tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:04 crc kubenswrapper[4770]: I1004 04:37:04.142446 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-784tz\" (UniqueName: \"kubernetes.io/projected/322194fd-446b-49b0-9051-20edd7db6d8a-kube-api-access-784tz\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:04 crc kubenswrapper[4770]: I1004 04:37:04.296143 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-js2xf" Oct 04 04:37:04 crc kubenswrapper[4770]: I1004 04:37:04.296171 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-js2xf" event={"ID":"2c3bbbec-3d17-487c-b37d-4ed7254d14f5","Type":"ContainerDied","Data":"613a8b5be18a31b3584486941632593a9873467269433e9733b038bf4de2ca4e"} Oct 04 04:37:04 crc kubenswrapper[4770]: I1004 04:37:04.296263 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="613a8b5be18a31b3584486941632593a9873467269433e9733b038bf4de2ca4e" Oct 04 04:37:04 crc kubenswrapper[4770]: I1004 04:37:04.299915 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cwx5p" event={"ID":"322194fd-446b-49b0-9051-20edd7db6d8a","Type":"ContainerDied","Data":"3455bda5d1f89b2083c36030413f632756df26d726e21bedd7bb9978473bfb7f"} Oct 04 04:37:04 crc kubenswrapper[4770]: I1004 04:37:04.300416 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3455bda5d1f89b2083c36030413f632756df26d726e21bedd7bb9978473bfb7f" Oct 04 04:37:04 crc kubenswrapper[4770]: I1004 04:37:04.300072 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cwx5p" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.433296 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-321b-account-create-q52jt"] Oct 04 04:37:09 crc kubenswrapper[4770]: E1004 04:37:09.434162 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e9b805f-4925-4291-9a15-a7efac711ba5" containerName="mariadb-database-create" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.434178 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e9b805f-4925-4291-9a15-a7efac711ba5" containerName="mariadb-database-create" Oct 04 04:37:09 crc kubenswrapper[4770]: E1004 04:37:09.434208 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c3bbbec-3d17-487c-b37d-4ed7254d14f5" containerName="mariadb-database-create" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.434218 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c3bbbec-3d17-487c-b37d-4ed7254d14f5" containerName="mariadb-database-create" Oct 04 04:37:09 crc kubenswrapper[4770]: E1004 04:37:09.434243 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322194fd-446b-49b0-9051-20edd7db6d8a" containerName="mariadb-database-create" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.434253 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="322194fd-446b-49b0-9051-20edd7db6d8a" containerName="mariadb-database-create" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.434461 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c3bbbec-3d17-487c-b37d-4ed7254d14f5" containerName="mariadb-database-create" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.434488 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e9b805f-4925-4291-9a15-a7efac711ba5" containerName="mariadb-database-create" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.434504 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="322194fd-446b-49b0-9051-20edd7db6d8a" containerName="mariadb-database-create" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.435228 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-321b-account-create-q52jt" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.438395 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.442232 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-321b-account-create-q52jt"] Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.636425 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sbxb\" (UniqueName: \"kubernetes.io/projected/b2b431c5-ac55-489e-ab00-bb9ee7176b77-kube-api-access-2sbxb\") pod \"nova-api-321b-account-create-q52jt\" (UID: \"b2b431c5-ac55-489e-ab00-bb9ee7176b77\") " pod="openstack/nova-api-321b-account-create-q52jt" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.638871 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-30a8-account-create-9sctc"] Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.640160 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30a8-account-create-9sctc" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.642140 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.657504 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-30a8-account-create-9sctc"] Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.738071 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sbxb\" (UniqueName: \"kubernetes.io/projected/b2b431c5-ac55-489e-ab00-bb9ee7176b77-kube-api-access-2sbxb\") pod \"nova-api-321b-account-create-q52jt\" (UID: \"b2b431c5-ac55-489e-ab00-bb9ee7176b77\") " pod="openstack/nova-api-321b-account-create-q52jt" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.761769 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sbxb\" (UniqueName: \"kubernetes.io/projected/b2b431c5-ac55-489e-ab00-bb9ee7176b77-kube-api-access-2sbxb\") pod \"nova-api-321b-account-create-q52jt\" (UID: \"b2b431c5-ac55-489e-ab00-bb9ee7176b77\") " pod="openstack/nova-api-321b-account-create-q52jt" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.767693 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-321b-account-create-q52jt" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.831832 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-44d4-account-create-c8pxc"] Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.833630 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-44d4-account-create-c8pxc" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.835950 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.840854 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptlnz\" (UniqueName: \"kubernetes.io/projected/66acc291-67a6-4e60-9c73-f9e7754853b7-kube-api-access-ptlnz\") pod \"nova-cell1-44d4-account-create-c8pxc\" (UID: \"66acc291-67a6-4e60-9c73-f9e7754853b7\") " pod="openstack/nova-cell1-44d4-account-create-c8pxc" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.840946 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9m59\" (UniqueName: \"kubernetes.io/projected/49759e6a-459f-4355-ab2f-fb45039cc611-kube-api-access-v9m59\") pod \"nova-cell0-30a8-account-create-9sctc\" (UID: \"49759e6a-459f-4355-ab2f-fb45039cc611\") " pod="openstack/nova-cell0-30a8-account-create-9sctc" Oct 04 04:37:09 crc kubenswrapper[4770]: I1004 04:37:09.847282 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-44d4-account-create-c8pxc"] Oct 04 04:37:10 crc kubenswrapper[4770]: I1004 04:37:09.943113 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9m59\" (UniqueName: \"kubernetes.io/projected/49759e6a-459f-4355-ab2f-fb45039cc611-kube-api-access-v9m59\") pod \"nova-cell0-30a8-account-create-9sctc\" (UID: \"49759e6a-459f-4355-ab2f-fb45039cc611\") " pod="openstack/nova-cell0-30a8-account-create-9sctc" Oct 04 04:37:10 crc kubenswrapper[4770]: I1004 04:37:09.943309 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptlnz\" (UniqueName: \"kubernetes.io/projected/66acc291-67a6-4e60-9c73-f9e7754853b7-kube-api-access-ptlnz\") pod \"nova-cell1-44d4-account-create-c8pxc\" (UID: \"66acc291-67a6-4e60-9c73-f9e7754853b7\") " pod="openstack/nova-cell1-44d4-account-create-c8pxc" Oct 04 04:37:10 crc kubenswrapper[4770]: I1004 04:37:09.962957 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9m59\" (UniqueName: \"kubernetes.io/projected/49759e6a-459f-4355-ab2f-fb45039cc611-kube-api-access-v9m59\") pod \"nova-cell0-30a8-account-create-9sctc\" (UID: \"49759e6a-459f-4355-ab2f-fb45039cc611\") " pod="openstack/nova-cell0-30a8-account-create-9sctc" Oct 04 04:37:10 crc kubenswrapper[4770]: I1004 04:37:09.966689 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptlnz\" (UniqueName: \"kubernetes.io/projected/66acc291-67a6-4e60-9c73-f9e7754853b7-kube-api-access-ptlnz\") pod \"nova-cell1-44d4-account-create-c8pxc\" (UID: \"66acc291-67a6-4e60-9c73-f9e7754853b7\") " pod="openstack/nova-cell1-44d4-account-create-c8pxc" Oct 04 04:37:10 crc kubenswrapper[4770]: I1004 04:37:10.010629 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30a8-account-create-9sctc" Oct 04 04:37:10 crc kubenswrapper[4770]: I1004 04:37:10.196442 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-44d4-account-create-c8pxc" Oct 04 04:37:11 crc kubenswrapper[4770]: I1004 04:37:11.193075 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-321b-account-create-q52jt"] Oct 04 04:37:11 crc kubenswrapper[4770]: I1004 04:37:11.206996 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 04:37:11 crc kubenswrapper[4770]: I1004 04:37:11.257016 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-30a8-account-create-9sctc"] Oct 04 04:37:11 crc kubenswrapper[4770]: I1004 04:37:11.263336 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-44d4-account-create-c8pxc"] Oct 04 04:37:11 crc kubenswrapper[4770]: W1004 04:37:11.277266 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66acc291_67a6_4e60_9c73_f9e7754853b7.slice/crio-f148e924a5401aad34d613c57e362352b27cf645e2fd648618b0a44958b2d147 WatchSource:0}: Error finding container f148e924a5401aad34d613c57e362352b27cf645e2fd648618b0a44958b2d147: Status 404 returned error can't find the container with id f148e924a5401aad34d613c57e362352b27cf645e2fd648618b0a44958b2d147 Oct 04 04:37:11 crc kubenswrapper[4770]: I1004 04:37:11.374089 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-321b-account-create-q52jt" event={"ID":"b2b431c5-ac55-489e-ab00-bb9ee7176b77","Type":"ContainerStarted","Data":"8a201fb82fd1c6c5f6a28b584958af29901f80b832f5143bd95d8b3331c3379d"} Oct 04 04:37:11 crc kubenswrapper[4770]: I1004 04:37:11.375400 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-44d4-account-create-c8pxc" event={"ID":"66acc291-67a6-4e60-9c73-f9e7754853b7","Type":"ContainerStarted","Data":"f148e924a5401aad34d613c57e362352b27cf645e2fd648618b0a44958b2d147"} Oct 04 04:37:11 crc kubenswrapper[4770]: I1004 04:37:11.376385 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30a8-account-create-9sctc" event={"ID":"49759e6a-459f-4355-ab2f-fb45039cc611","Type":"ContainerStarted","Data":"571e404a5ad46039f1e47a6f69d8333be91cf2f104f2e2182066bd894c492c49"} Oct 04 04:37:12 crc kubenswrapper[4770]: I1004 04:37:12.390965 4770 generic.go:334] "Generic (PLEG): container finished" podID="b2b431c5-ac55-489e-ab00-bb9ee7176b77" containerID="1cb5eb88ff8f2484b1899265fafde1b2fce532ca1ea7bcba7c3cc23972f96d60" exitCode=0 Oct 04 04:37:12 crc kubenswrapper[4770]: I1004 04:37:12.391061 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-321b-account-create-q52jt" event={"ID":"b2b431c5-ac55-489e-ab00-bb9ee7176b77","Type":"ContainerDied","Data":"1cb5eb88ff8f2484b1899265fafde1b2fce532ca1ea7bcba7c3cc23972f96d60"} Oct 04 04:37:12 crc kubenswrapper[4770]: I1004 04:37:12.393857 4770 generic.go:334] "Generic (PLEG): container finished" podID="66acc291-67a6-4e60-9c73-f9e7754853b7" containerID="21a22851314d295d0c1bad2cc6a6b344fbf90d7e2a3f187db45faf02acadc4ab" exitCode=0 Oct 04 04:37:12 crc kubenswrapper[4770]: I1004 04:37:12.393981 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-44d4-account-create-c8pxc" event={"ID":"66acc291-67a6-4e60-9c73-f9e7754853b7","Type":"ContainerDied","Data":"21a22851314d295d0c1bad2cc6a6b344fbf90d7e2a3f187db45faf02acadc4ab"} Oct 04 04:37:12 crc kubenswrapper[4770]: I1004 04:37:12.400716 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30a8-account-create-9sctc" event={"ID":"49759e6a-459f-4355-ab2f-fb45039cc611","Type":"ContainerStarted","Data":"9d38ba5616e3dc64bcc6918aa860e976782fff527fa61b4c5b028401c0739d8f"} Oct 04 04:37:12 crc kubenswrapper[4770]: I1004 04:37:12.441397 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-30a8-account-create-9sctc" podStartSLOduration=3.441370457 podStartE2EDuration="3.441370457s" podCreationTimestamp="2025-10-04 04:37:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:12.43573762 +0000 UTC m=+5643.727747352" watchObservedRunningTime="2025-10-04 04:37:12.441370457 +0000 UTC m=+5643.733380169" Oct 04 04:37:13 crc kubenswrapper[4770]: I1004 04:37:13.411242 4770 generic.go:334] "Generic (PLEG): container finished" podID="49759e6a-459f-4355-ab2f-fb45039cc611" containerID="9d38ba5616e3dc64bcc6918aa860e976782fff527fa61b4c5b028401c0739d8f" exitCode=0 Oct 04 04:37:13 crc kubenswrapper[4770]: I1004 04:37:13.411297 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30a8-account-create-9sctc" event={"ID":"49759e6a-459f-4355-ab2f-fb45039cc611","Type":"ContainerDied","Data":"9d38ba5616e3dc64bcc6918aa860e976782fff527fa61b4c5b028401c0739d8f"} Oct 04 04:37:13 crc kubenswrapper[4770]: I1004 04:37:13.841937 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-321b-account-create-q52jt" Oct 04 04:37:13 crc kubenswrapper[4770]: I1004 04:37:13.848435 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-44d4-account-create-c8pxc" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.013257 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptlnz\" (UniqueName: \"kubernetes.io/projected/66acc291-67a6-4e60-9c73-f9e7754853b7-kube-api-access-ptlnz\") pod \"66acc291-67a6-4e60-9c73-f9e7754853b7\" (UID: \"66acc291-67a6-4e60-9c73-f9e7754853b7\") " Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.013665 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sbxb\" (UniqueName: \"kubernetes.io/projected/b2b431c5-ac55-489e-ab00-bb9ee7176b77-kube-api-access-2sbxb\") pod \"b2b431c5-ac55-489e-ab00-bb9ee7176b77\" (UID: \"b2b431c5-ac55-489e-ab00-bb9ee7176b77\") " Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.020199 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66acc291-67a6-4e60-9c73-f9e7754853b7-kube-api-access-ptlnz" (OuterVolumeSpecName: "kube-api-access-ptlnz") pod "66acc291-67a6-4e60-9c73-f9e7754853b7" (UID: "66acc291-67a6-4e60-9c73-f9e7754853b7"). InnerVolumeSpecName "kube-api-access-ptlnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.021091 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b431c5-ac55-489e-ab00-bb9ee7176b77-kube-api-access-2sbxb" (OuterVolumeSpecName: "kube-api-access-2sbxb") pod "b2b431c5-ac55-489e-ab00-bb9ee7176b77" (UID: "b2b431c5-ac55-489e-ab00-bb9ee7176b77"). InnerVolumeSpecName "kube-api-access-2sbxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.115891 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sbxb\" (UniqueName: \"kubernetes.io/projected/b2b431c5-ac55-489e-ab00-bb9ee7176b77-kube-api-access-2sbxb\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.115943 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptlnz\" (UniqueName: \"kubernetes.io/projected/66acc291-67a6-4e60-9c73-f9e7754853b7-kube-api-access-ptlnz\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.423390 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-321b-account-create-q52jt" event={"ID":"b2b431c5-ac55-489e-ab00-bb9ee7176b77","Type":"ContainerDied","Data":"8a201fb82fd1c6c5f6a28b584958af29901f80b832f5143bd95d8b3331c3379d"} Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.423436 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-321b-account-create-q52jt" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.423448 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a201fb82fd1c6c5f6a28b584958af29901f80b832f5143bd95d8b3331c3379d" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.426160 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-44d4-account-create-c8pxc" event={"ID":"66acc291-67a6-4e60-9c73-f9e7754853b7","Type":"ContainerDied","Data":"f148e924a5401aad34d613c57e362352b27cf645e2fd648618b0a44958b2d147"} Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.426219 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-44d4-account-create-c8pxc" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.426222 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f148e924a5401aad34d613c57e362352b27cf645e2fd648618b0a44958b2d147" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.799341 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30a8-account-create-9sctc" Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.938273 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9m59\" (UniqueName: \"kubernetes.io/projected/49759e6a-459f-4355-ab2f-fb45039cc611-kube-api-access-v9m59\") pod \"49759e6a-459f-4355-ab2f-fb45039cc611\" (UID: \"49759e6a-459f-4355-ab2f-fb45039cc611\") " Oct 04 04:37:14 crc kubenswrapper[4770]: I1004 04:37:14.944950 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49759e6a-459f-4355-ab2f-fb45039cc611-kube-api-access-v9m59" (OuterVolumeSpecName: "kube-api-access-v9m59") pod "49759e6a-459f-4355-ab2f-fb45039cc611" (UID: "49759e6a-459f-4355-ab2f-fb45039cc611"). InnerVolumeSpecName "kube-api-access-v9m59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:15 crc kubenswrapper[4770]: I1004 04:37:15.041504 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9m59\" (UniqueName: \"kubernetes.io/projected/49759e6a-459f-4355-ab2f-fb45039cc611-kube-api-access-v9m59\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:15 crc kubenswrapper[4770]: I1004 04:37:15.436818 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30a8-account-create-9sctc" event={"ID":"49759e6a-459f-4355-ab2f-fb45039cc611","Type":"ContainerDied","Data":"571e404a5ad46039f1e47a6f69d8333be91cf2f104f2e2182066bd894c492c49"} Oct 04 04:37:15 crc kubenswrapper[4770]: I1004 04:37:15.436863 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="571e404a5ad46039f1e47a6f69d8333be91cf2f104f2e2182066bd894c492c49" Oct 04 04:37:15 crc kubenswrapper[4770]: I1004 04:37:15.436896 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30a8-account-create-9sctc" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.982146 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kd4rs"] Oct 04 04:37:19 crc kubenswrapper[4770]: E1004 04:37:19.982506 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49759e6a-459f-4355-ab2f-fb45039cc611" containerName="mariadb-account-create" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.982518 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="49759e6a-459f-4355-ab2f-fb45039cc611" containerName="mariadb-account-create" Oct 04 04:37:19 crc kubenswrapper[4770]: E1004 04:37:19.982532 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66acc291-67a6-4e60-9c73-f9e7754853b7" containerName="mariadb-account-create" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.982538 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="66acc291-67a6-4e60-9c73-f9e7754853b7" containerName="mariadb-account-create" Oct 04 04:37:19 crc kubenswrapper[4770]: E1004 04:37:19.982561 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b431c5-ac55-489e-ab00-bb9ee7176b77" containerName="mariadb-account-create" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.982568 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b431c5-ac55-489e-ab00-bb9ee7176b77" containerName="mariadb-account-create" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.982746 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="49759e6a-459f-4355-ab2f-fb45039cc611" containerName="mariadb-account-create" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.982766 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2b431c5-ac55-489e-ab00-bb9ee7176b77" containerName="mariadb-account-create" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.982780 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="66acc291-67a6-4e60-9c73-f9e7754853b7" containerName="mariadb-account-create" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.983336 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.985664 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fclmn" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.985711 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.991989 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 04 04:37:19 crc kubenswrapper[4770]: I1004 04:37:19.995561 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kd4rs"] Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.140765 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-scripts\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.140840 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.140919 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6mmn\" (UniqueName: \"kubernetes.io/projected/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-kube-api-access-t6mmn\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.141127 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-config-data\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.242670 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-scripts\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.242759 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.242814 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6mmn\" (UniqueName: \"kubernetes.io/projected/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-kube-api-access-t6mmn\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.242850 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-config-data\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.248493 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.249461 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-scripts\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.250662 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-config-data\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.260673 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6mmn\" (UniqueName: \"kubernetes.io/projected/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-kube-api-access-t6mmn\") pod \"nova-cell0-conductor-db-sync-kd4rs\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.311630 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:20 crc kubenswrapper[4770]: I1004 04:37:20.787207 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kd4rs"] Oct 04 04:37:20 crc kubenswrapper[4770]: W1004 04:37:20.792512 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0e86553_7a36_44f8_b8fb_e21de6d2bda0.slice/crio-31d53b84d4f2230bb34628d2f63767a5edfc1c9b0e7cfccb778d4723b16fe5ba WatchSource:0}: Error finding container 31d53b84d4f2230bb34628d2f63767a5edfc1c9b0e7cfccb778d4723b16fe5ba: Status 404 returned error can't find the container with id 31d53b84d4f2230bb34628d2f63767a5edfc1c9b0e7cfccb778d4723b16fe5ba Oct 04 04:37:21 crc kubenswrapper[4770]: I1004 04:37:21.503594 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kd4rs" event={"ID":"b0e86553-7a36-44f8-b8fb-e21de6d2bda0","Type":"ContainerStarted","Data":"cfde03b252b23508a80eee27e883f8579b863b9a2243294ca7acf7dd8f0b9717"} Oct 04 04:37:21 crc kubenswrapper[4770]: I1004 04:37:21.504329 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kd4rs" event={"ID":"b0e86553-7a36-44f8-b8fb-e21de6d2bda0","Type":"ContainerStarted","Data":"31d53b84d4f2230bb34628d2f63767a5edfc1c9b0e7cfccb778d4723b16fe5ba"} Oct 04 04:37:21 crc kubenswrapper[4770]: I1004 04:37:21.523206 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-kd4rs" podStartSLOduration=2.523178738 podStartE2EDuration="2.523178738s" podCreationTimestamp="2025-10-04 04:37:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:21.52169996 +0000 UTC m=+5652.813709692" watchObservedRunningTime="2025-10-04 04:37:21.523178738 +0000 UTC m=+5652.815188470" Oct 04 04:37:30 crc kubenswrapper[4770]: I1004 04:37:30.610205 4770 generic.go:334] "Generic (PLEG): container finished" podID="b0e86553-7a36-44f8-b8fb-e21de6d2bda0" containerID="cfde03b252b23508a80eee27e883f8579b863b9a2243294ca7acf7dd8f0b9717" exitCode=0 Oct 04 04:37:30 crc kubenswrapper[4770]: I1004 04:37:30.610398 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kd4rs" event={"ID":"b0e86553-7a36-44f8-b8fb-e21de6d2bda0","Type":"ContainerDied","Data":"cfde03b252b23508a80eee27e883f8579b863b9a2243294ca7acf7dd8f0b9717"} Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.051848 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8g92n"] Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.054768 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.068328 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8g92n"] Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.085321 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-catalog-content\") pod \"redhat-marketplace-8g92n\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.085387 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-utilities\") pod \"redhat-marketplace-8g92n\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.085457 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrwx4\" (UniqueName: \"kubernetes.io/projected/92c61d37-84ca-4f8b-9911-5d1ad7c05074-kube-api-access-hrwx4\") pod \"redhat-marketplace-8g92n\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.187618 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-catalog-content\") pod \"redhat-marketplace-8g92n\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.187672 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-utilities\") pod \"redhat-marketplace-8g92n\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.187747 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrwx4\" (UniqueName: \"kubernetes.io/projected/92c61d37-84ca-4f8b-9911-5d1ad7c05074-kube-api-access-hrwx4\") pod \"redhat-marketplace-8g92n\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.188589 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-catalog-content\") pod \"redhat-marketplace-8g92n\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.188676 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-utilities\") pod \"redhat-marketplace-8g92n\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.209966 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrwx4\" (UniqueName: \"kubernetes.io/projected/92c61d37-84ca-4f8b-9911-5d1ad7c05074-kube-api-access-hrwx4\") pod \"redhat-marketplace-8g92n\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.386887 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.813213 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8g92n"] Oct 04 04:37:31 crc kubenswrapper[4770]: I1004 04:37:31.977574 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.002701 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-scripts\") pod \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.002873 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-config-data\") pod \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.003047 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-combined-ca-bundle\") pod \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.003141 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6mmn\" (UniqueName: \"kubernetes.io/projected/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-kube-api-access-t6mmn\") pod \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\" (UID: \"b0e86553-7a36-44f8-b8fb-e21de6d2bda0\") " Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.010349 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-kube-api-access-t6mmn" (OuterVolumeSpecName: "kube-api-access-t6mmn") pod "b0e86553-7a36-44f8-b8fb-e21de6d2bda0" (UID: "b0e86553-7a36-44f8-b8fb-e21de6d2bda0"). InnerVolumeSpecName "kube-api-access-t6mmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.010504 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-scripts" (OuterVolumeSpecName: "scripts") pod "b0e86553-7a36-44f8-b8fb-e21de6d2bda0" (UID: "b0e86553-7a36-44f8-b8fb-e21de6d2bda0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.051241 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0e86553-7a36-44f8-b8fb-e21de6d2bda0" (UID: "b0e86553-7a36-44f8-b8fb-e21de6d2bda0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.065168 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-config-data" (OuterVolumeSpecName: "config-data") pod "b0e86553-7a36-44f8-b8fb-e21de6d2bda0" (UID: "b0e86553-7a36-44f8-b8fb-e21de6d2bda0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.105079 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.105114 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6mmn\" (UniqueName: \"kubernetes.io/projected/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-kube-api-access-t6mmn\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.105130 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.105141 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e86553-7a36-44f8-b8fb-e21de6d2bda0-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.637418 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kd4rs" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.637444 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kd4rs" event={"ID":"b0e86553-7a36-44f8-b8fb-e21de6d2bda0","Type":"ContainerDied","Data":"31d53b84d4f2230bb34628d2f63767a5edfc1c9b0e7cfccb778d4723b16fe5ba"} Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.637582 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31d53b84d4f2230bb34628d2f63767a5edfc1c9b0e7cfccb778d4723b16fe5ba" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.640419 4770 generic.go:334] "Generic (PLEG): container finished" podID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerID="91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60" exitCode=0 Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.640511 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8g92n" event={"ID":"92c61d37-84ca-4f8b-9911-5d1ad7c05074","Type":"ContainerDied","Data":"91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60"} Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.640564 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8g92n" event={"ID":"92c61d37-84ca-4f8b-9911-5d1ad7c05074","Type":"ContainerStarted","Data":"6826dc9e206983977198bb4853599c064fc9252a125f83a824cb9f7a6da28be6"} Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.643520 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.767956 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:37:32 crc kubenswrapper[4770]: E1004 04:37:32.768543 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e86553-7a36-44f8-b8fb-e21de6d2bda0" containerName="nova-cell0-conductor-db-sync" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.768607 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e86553-7a36-44f8-b8fb-e21de6d2bda0" containerName="nova-cell0-conductor-db-sync" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.768832 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0e86553-7a36-44f8-b8fb-e21de6d2bda0" containerName="nova-cell0-conductor-db-sync" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.769483 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.773047 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.774740 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fclmn" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.795509 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.820715 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.821705 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdndj\" (UniqueName: \"kubernetes.io/projected/048a0b66-f6e3-424c-80c5-552afd14ed64-kube-api-access-hdndj\") pod \"nova-cell0-conductor-0\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.821804 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.923182 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdndj\" (UniqueName: \"kubernetes.io/projected/048a0b66-f6e3-424c-80c5-552afd14ed64-kube-api-access-hdndj\") pod \"nova-cell0-conductor-0\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.923237 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:32 crc kubenswrapper[4770]: I1004 04:37:32.923307 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:33 crc kubenswrapper[4770]: I1004 04:37:33.175734 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:33 crc kubenswrapper[4770]: I1004 04:37:33.178504 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:33 crc kubenswrapper[4770]: I1004 04:37:33.283171 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdndj\" (UniqueName: \"kubernetes.io/projected/048a0b66-f6e3-424c-80c5-552afd14ed64-kube-api-access-hdndj\") pod \"nova-cell0-conductor-0\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:33 crc kubenswrapper[4770]: I1004 04:37:33.402308 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:33 crc kubenswrapper[4770]: I1004 04:37:33.889575 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:37:34 crc kubenswrapper[4770]: I1004 04:37:34.687672 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"048a0b66-f6e3-424c-80c5-552afd14ed64","Type":"ContainerStarted","Data":"d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b"} Oct 04 04:37:34 crc kubenswrapper[4770]: I1004 04:37:34.688052 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"048a0b66-f6e3-424c-80c5-552afd14ed64","Type":"ContainerStarted","Data":"3c5ce60c052f33b56d736d2c60294911bed95f56d15c06d8d62b5a1bbe496be8"} Oct 04 04:37:34 crc kubenswrapper[4770]: I1004 04:37:34.688086 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:34 crc kubenswrapper[4770]: I1004 04:37:34.716202 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.71615757 podStartE2EDuration="2.71615757s" podCreationTimestamp="2025-10-04 04:37:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:34.708285924 +0000 UTC m=+5666.000295676" watchObservedRunningTime="2025-10-04 04:37:34.71615757 +0000 UTC m=+5666.008167292" Oct 04 04:37:35 crc kubenswrapper[4770]: I1004 04:37:35.710720 4770 generic.go:334] "Generic (PLEG): container finished" podID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerID="ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c" exitCode=0 Oct 04 04:37:35 crc kubenswrapper[4770]: I1004 04:37:35.710788 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8g92n" event={"ID":"92c61d37-84ca-4f8b-9911-5d1ad7c05074","Type":"ContainerDied","Data":"ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c"} Oct 04 04:37:36 crc kubenswrapper[4770]: I1004 04:37:36.740807 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8g92n" event={"ID":"92c61d37-84ca-4f8b-9911-5d1ad7c05074","Type":"ContainerStarted","Data":"a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864"} Oct 04 04:37:36 crc kubenswrapper[4770]: I1004 04:37:36.760365 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8g92n" podStartSLOduration=2.017024182 podStartE2EDuration="5.760347777s" podCreationTimestamp="2025-10-04 04:37:31 +0000 UTC" firstStartedPulling="2025-10-04 04:37:32.643331345 +0000 UTC m=+5663.935341057" lastFinishedPulling="2025-10-04 04:37:36.38665494 +0000 UTC m=+5667.678664652" observedRunningTime="2025-10-04 04:37:36.758814636 +0000 UTC m=+5668.050824368" watchObservedRunningTime="2025-10-04 04:37:36.760347777 +0000 UTC m=+5668.052357499" Oct 04 04:37:41 crc kubenswrapper[4770]: I1004 04:37:41.388880 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:41 crc kubenswrapper[4770]: I1004 04:37:41.393274 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:41 crc kubenswrapper[4770]: I1004 04:37:41.438643 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:41 crc kubenswrapper[4770]: I1004 04:37:41.845981 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:41 crc kubenswrapper[4770]: I1004 04:37:41.901975 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8g92n"] Oct 04 04:37:43 crc kubenswrapper[4770]: I1004 04:37:43.435922 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 04:37:43 crc kubenswrapper[4770]: I1004 04:37:43.812705 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8g92n" podUID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerName="registry-server" containerID="cri-o://a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864" gracePeriod=2 Oct 04 04:37:43 crc kubenswrapper[4770]: I1004 04:37:43.939060 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-gpw5q"] Oct 04 04:37:43 crc kubenswrapper[4770]: I1004 04:37:43.940484 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:43 crc kubenswrapper[4770]: I1004 04:37:43.945168 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 04 04:37:43 crc kubenswrapper[4770]: I1004 04:37:43.945502 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 04 04:37:43 crc kubenswrapper[4770]: I1004 04:37:43.958678 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gpw5q"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.037234 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thwnw\" (UniqueName: \"kubernetes.io/projected/38e50e99-f55d-47a6-ab20-fc4c37c91852-kube-api-access-thwnw\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.037646 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-scripts\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.037722 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.037781 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-config-data\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.098203 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.100441 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.102954 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.119700 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.139683 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.139751 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9xxw\" (UniqueName: \"kubernetes.io/projected/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-kube-api-access-w9xxw\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.139819 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-config-data\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.139917 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.139950 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thwnw\" (UniqueName: \"kubernetes.io/projected/38e50e99-f55d-47a6-ab20-fc4c37c91852-kube-api-access-thwnw\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.139974 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-logs\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.140043 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-config-data\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.140084 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-scripts\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.157453 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-config-data\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.160115 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-scripts\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.170248 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.180474 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.181702 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.188355 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.189756 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thwnw\" (UniqueName: \"kubernetes.io/projected/38e50e99-f55d-47a6-ab20-fc4c37c91852-kube-api-access-thwnw\") pod \"nova-cell0-cell-mapping-gpw5q\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.231247 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.234333 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.236056 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.247968 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-config-data\") pod \"nova-scheduler-0\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.248196 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9xxw\" (UniqueName: \"kubernetes.io/projected/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-kube-api-access-w9xxw\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.248394 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.248425 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqzgw\" (UniqueName: \"kubernetes.io/projected/499c7c37-0517-4c0d-94cd-f0af00466cef-kube-api-access-bqzgw\") pod \"nova-scheduler-0\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.248473 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.248508 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-logs\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.248588 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-config-data\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.255634 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-logs\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.273614 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-config-data\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.298480 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9xxw\" (UniqueName: \"kubernetes.io/projected/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-kube-api-access-w9xxw\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.299037 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.300676 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.332550 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.352914 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-config-data\") pod \"nova-scheduler-0\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.352956 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.353014 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f51bb4-55de-41b6-a91a-9b7b965ac20c-logs\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.353038 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v6g5\" (UniqueName: \"kubernetes.io/projected/17f51bb4-55de-41b6-a91a-9b7b965ac20c-kube-api-access-9v6g5\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.356078 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.356113 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqzgw\" (UniqueName: \"kubernetes.io/projected/499c7c37-0517-4c0d-94cd-f0af00466cef-kube-api-access-bqzgw\") pod \"nova-scheduler-0\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.356146 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-config-data\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.356729 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-config-data\") pod \"nova-scheduler-0\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.358961 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.370889 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.379457 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqzgw\" (UniqueName: \"kubernetes.io/projected/499c7c37-0517-4c0d-94cd-f0af00466cef-kube-api-access-bqzgw\") pod \"nova-scheduler-0\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.400891 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.402548 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.410908 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.411850 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.422583 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.431445 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7784f5b875-slvnh"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.440982 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.443916 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7784f5b875-slvnh"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.457635 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.457710 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-config-data\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.457842 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.457910 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.457959 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f51bb4-55de-41b6-a91a-9b7b965ac20c-logs\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.457977 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/39e02f97-9bb3-43b4-a2f8-7633252519a3-kube-api-access-hbjp4\") pod \"nova-cell1-novncproxy-0\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.458014 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v6g5\" (UniqueName: \"kubernetes.io/projected/17f51bb4-55de-41b6-a91a-9b7b965ac20c-kube-api-access-9v6g5\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.458931 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f51bb4-55de-41b6-a91a-9b7b965ac20c-logs\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.473851 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-config-data\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.473856 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.477789 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v6g5\" (UniqueName: \"kubernetes.io/projected/17f51bb4-55de-41b6-a91a-9b7b965ac20c-kube-api-access-9v6g5\") pod \"nova-metadata-0\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.517288 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.557044 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.558752 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrwx4\" (UniqueName: \"kubernetes.io/projected/92c61d37-84ca-4f8b-9911-5d1ad7c05074-kube-api-access-hrwx4\") pod \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.558805 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-utilities\") pod \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.558872 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-catalog-content\") pod \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\" (UID: \"92c61d37-84ca-4f8b-9911-5d1ad7c05074\") " Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.559295 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.559335 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-nb\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.559375 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.559391 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-dns-svc\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.559450 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/39e02f97-9bb3-43b4-a2f8-7633252519a3-kube-api-access-hbjp4\") pod \"nova-cell1-novncproxy-0\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.559480 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-sb\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.559507 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk447\" (UniqueName: \"kubernetes.io/projected/facf4f26-1f71-4f73-9d66-aaeb7476628d-kube-api-access-nk447\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.559524 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-config\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.560652 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-utilities" (OuterVolumeSpecName: "utilities") pod "92c61d37-84ca-4f8b-9911-5d1ad7c05074" (UID: "92c61d37-84ca-4f8b-9911-5d1ad7c05074"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.567849 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.577531 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.577750 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92c61d37-84ca-4f8b-9911-5d1ad7c05074-kube-api-access-hrwx4" (OuterVolumeSpecName: "kube-api-access-hrwx4") pod "92c61d37-84ca-4f8b-9911-5d1ad7c05074" (UID: "92c61d37-84ca-4f8b-9911-5d1ad7c05074"). InnerVolumeSpecName "kube-api-access-hrwx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.580122 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92c61d37-84ca-4f8b-9911-5d1ad7c05074" (UID: "92c61d37-84ca-4f8b-9911-5d1ad7c05074"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.586106 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/39e02f97-9bb3-43b4-a2f8-7633252519a3-kube-api-access-hbjp4\") pod \"nova-cell1-novncproxy-0\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.658409 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.660856 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-nb\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.660927 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-dns-svc\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.660999 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-sb\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.661067 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk447\" (UniqueName: \"kubernetes.io/projected/facf4f26-1f71-4f73-9d66-aaeb7476628d-kube-api-access-nk447\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.661088 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-config\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.661189 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.661204 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c61d37-84ca-4f8b-9911-5d1ad7c05074-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.661218 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrwx4\" (UniqueName: \"kubernetes.io/projected/92c61d37-84ca-4f8b-9911-5d1ad7c05074-kube-api-access-hrwx4\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.662241 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-config\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.662896 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-nb\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.663044 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-sb\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.663745 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-dns-svc\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.694652 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk447\" (UniqueName: \"kubernetes.io/projected/facf4f26-1f71-4f73-9d66-aaeb7476628d-kube-api-access-nk447\") pod \"dnsmasq-dns-7784f5b875-slvnh\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.729494 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.766430 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.838264 4770 generic.go:334] "Generic (PLEG): container finished" podID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerID="a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864" exitCode=0 Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.838327 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8g92n" event={"ID":"92c61d37-84ca-4f8b-9911-5d1ad7c05074","Type":"ContainerDied","Data":"a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864"} Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.838358 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8g92n" event={"ID":"92c61d37-84ca-4f8b-9911-5d1ad7c05074","Type":"ContainerDied","Data":"6826dc9e206983977198bb4853599c064fc9252a125f83a824cb9f7a6da28be6"} Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.838377 4770 scope.go:117] "RemoveContainer" containerID="a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.838555 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8g92n" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.910064 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gpw5q"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.912328 4770 scope.go:117] "RemoveContainer" containerID="ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c" Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.920434 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8g92n"] Oct 04 04:37:44 crc kubenswrapper[4770]: I1004 04:37:44.929736 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8g92n"] Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.058198 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:37:45 crc kubenswrapper[4770]: W1004 04:37:45.062543 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecc5739f_51fe_4aee_a31f_2c8654fd9edf.slice/crio-7492a9a1a4be09a52546603d2bea783742b42b146430da37bfcac7e70eed1bcf WatchSource:0}: Error finding container 7492a9a1a4be09a52546603d2bea783742b42b146430da37bfcac7e70eed1bcf: Status 404 returned error can't find the container with id 7492a9a1a4be09a52546603d2bea783742b42b146430da37bfcac7e70eed1bcf Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.080412 4770 scope.go:117] "RemoveContainer" containerID="91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.151815 4770 scope.go:117] "RemoveContainer" containerID="a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864" Oct 04 04:37:45 crc kubenswrapper[4770]: E1004 04:37:45.152328 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864\": container with ID starting with a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864 not found: ID does not exist" containerID="a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.152362 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864"} err="failed to get container status \"a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864\": rpc error: code = NotFound desc = could not find container \"a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864\": container with ID starting with a5fb9dedaf35bbdb4e7bfb569e7723993a046abc3e89cd12aacada734457a864 not found: ID does not exist" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.152388 4770 scope.go:117] "RemoveContainer" containerID="ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c" Oct 04 04:37:45 crc kubenswrapper[4770]: E1004 04:37:45.152668 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c\": container with ID starting with ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c not found: ID does not exist" containerID="ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.152697 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c"} err="failed to get container status \"ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c\": rpc error: code = NotFound desc = could not find container \"ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c\": container with ID starting with ea7dba7a6f46137b65632ba858674d511b1e2528a42bc59f8422286b0a55074c not found: ID does not exist" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.152715 4770 scope.go:117] "RemoveContainer" containerID="91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60" Oct 04 04:37:45 crc kubenswrapper[4770]: E1004 04:37:45.153128 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60\": container with ID starting with 91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60 not found: ID does not exist" containerID="91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.153157 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60"} err="failed to get container status \"91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60\": rpc error: code = NotFound desc = could not find container \"91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60\": container with ID starting with 91ad12ab5adbbc4ce88eda1d3d5d2377d4dcd2f907f466bf8f78f69ba5be7c60 not found: ID does not exist" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.257609 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:37:45 crc kubenswrapper[4770]: W1004 04:37:45.263683 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod499c7c37_0517_4c0d_94cd_f0af00466cef.slice/crio-cf27ea9914999e75a81ff2b1a04d9c0c503692018964ff91a9947aabee98763e WatchSource:0}: Error finding container cf27ea9914999e75a81ff2b1a04d9c0c503692018964ff91a9947aabee98763e: Status 404 returned error can't find the container with id cf27ea9914999e75a81ff2b1a04d9c0c503692018964ff91a9947aabee98763e Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.466871 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.477341 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:37:45 crc kubenswrapper[4770]: W1004 04:37:45.478578 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17f51bb4_55de_41b6_a91a_9b7b965ac20c.slice/crio-385cb8d729bc08b6ffccc9beaf380a8413f56a40701225f9a87346f280fdde1f WatchSource:0}: Error finding container 385cb8d729bc08b6ffccc9beaf380a8413f56a40701225f9a87346f280fdde1f: Status 404 returned error can't find the container with id 385cb8d729bc08b6ffccc9beaf380a8413f56a40701225f9a87346f280fdde1f Oct 04 04:37:45 crc kubenswrapper[4770]: W1004 04:37:45.489218 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39e02f97_9bb3_43b4_a2f8_7633252519a3.slice/crio-2292442289fe31fec881436081a7ccc3450c73e61cc4c800213cfa4e04ee8caa WatchSource:0}: Error finding container 2292442289fe31fec881436081a7ccc3450c73e61cc4c800213cfa4e04ee8caa: Status 404 returned error can't find the container with id 2292442289fe31fec881436081a7ccc3450c73e61cc4c800213cfa4e04ee8caa Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.494157 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7784f5b875-slvnh"] Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.564629 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdzqf"] Oct 04 04:37:45 crc kubenswrapper[4770]: E1004 04:37:45.565102 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerName="extract-utilities" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.565124 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerName="extract-utilities" Oct 04 04:37:45 crc kubenswrapper[4770]: E1004 04:37:45.565147 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerName="extract-content" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.565156 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerName="extract-content" Oct 04 04:37:45 crc kubenswrapper[4770]: E1004 04:37:45.565177 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerName="registry-server" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.565186 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerName="registry-server" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.565400 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" containerName="registry-server" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.567331 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.570842 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.570991 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.571777 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdzqf"] Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.584403 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-config-data\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.584453 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-scripts\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.584484 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.584563 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b25r\" (UniqueName: \"kubernetes.io/projected/3c093be9-71c5-4478-84b2-da4d87396fc9-kube-api-access-6b25r\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.687166 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-config-data\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.687669 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-scripts\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.687702 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.687777 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b25r\" (UniqueName: \"kubernetes.io/projected/3c093be9-71c5-4478-84b2-da4d87396fc9-kube-api-access-6b25r\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.688229 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92c61d37-84ca-4f8b-9911-5d1ad7c05074" path="/var/lib/kubelet/pods/92c61d37-84ca-4f8b-9911-5d1ad7c05074/volumes" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.691195 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-config-data\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.691743 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.692099 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-scripts\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.705488 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b25r\" (UniqueName: \"kubernetes.io/projected/3c093be9-71c5-4478-84b2-da4d87396fc9-kube-api-access-6b25r\") pod \"nova-cell1-conductor-db-sync-xdzqf\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.849285 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f51bb4-55de-41b6-a91a-9b7b965ac20c","Type":"ContainerStarted","Data":"7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.849327 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f51bb4-55de-41b6-a91a-9b7b965ac20c","Type":"ContainerStarted","Data":"385cb8d729bc08b6ffccc9beaf380a8413f56a40701225f9a87346f280fdde1f"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.851163 4770 generic.go:334] "Generic (PLEG): container finished" podID="facf4f26-1f71-4f73-9d66-aaeb7476628d" containerID="92171d8dbe7b6e0888d2f9ad95bbc962b7fb273c5a8868260336ae6e5d234b29" exitCode=0 Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.851233 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" event={"ID":"facf4f26-1f71-4f73-9d66-aaeb7476628d","Type":"ContainerDied","Data":"92171d8dbe7b6e0888d2f9ad95bbc962b7fb273c5a8868260336ae6e5d234b29"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.851261 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" event={"ID":"facf4f26-1f71-4f73-9d66-aaeb7476628d","Type":"ContainerStarted","Data":"1fcbbda759a57d98340482a6357a339c96b1edee92ae7aca954a69a8de13753f"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.852841 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gpw5q" event={"ID":"38e50e99-f55d-47a6-ab20-fc4c37c91852","Type":"ContainerStarted","Data":"abb7af7395e58bb4795f53134cc7c15f7698cbe539b0c93e346617e09db55e5c"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.852869 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gpw5q" event={"ID":"38e50e99-f55d-47a6-ab20-fc4c37c91852","Type":"ContainerStarted","Data":"007217b67c6ff7295f0c24ecedb69e0d8575642a23b04c8518823e0038df8e60"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.856872 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ecc5739f-51fe-4aee-a31f-2c8654fd9edf","Type":"ContainerStarted","Data":"3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.856904 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ecc5739f-51fe-4aee-a31f-2c8654fd9edf","Type":"ContainerStarted","Data":"bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.856914 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ecc5739f-51fe-4aee-a31f-2c8654fd9edf","Type":"ContainerStarted","Data":"7492a9a1a4be09a52546603d2bea783742b42b146430da37bfcac7e70eed1bcf"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.864123 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"499c7c37-0517-4c0d-94cd-f0af00466cef","Type":"ContainerStarted","Data":"1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.864169 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"499c7c37-0517-4c0d-94cd-f0af00466cef","Type":"ContainerStarted","Data":"cf27ea9914999e75a81ff2b1a04d9c0c503692018964ff91a9947aabee98763e"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.880055 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"39e02f97-9bb3-43b4-a2f8-7633252519a3","Type":"ContainerStarted","Data":"1513f37950192789203ff20cde5e658cc3e08a222bcc4798c8e0bf256043e3e5"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.880100 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"39e02f97-9bb3-43b4-a2f8-7633252519a3","Type":"ContainerStarted","Data":"2292442289fe31fec881436081a7ccc3450c73e61cc4c800213cfa4e04ee8caa"} Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.907517 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.910481 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-gpw5q" podStartSLOduration=2.910460325 podStartE2EDuration="2.910460325s" podCreationTimestamp="2025-10-04 04:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:45.888944162 +0000 UTC m=+5677.180953874" watchObservedRunningTime="2025-10-04 04:37:45.910460325 +0000 UTC m=+5677.202470037" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.954593 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.954575057 podStartE2EDuration="1.954575057s" podCreationTimestamp="2025-10-04 04:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:45.906586033 +0000 UTC m=+5677.198595745" watchObservedRunningTime="2025-10-04 04:37:45.954575057 +0000 UTC m=+5677.246584769" Oct 04 04:37:45 crc kubenswrapper[4770]: I1004 04:37:45.975460 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.975415202 podStartE2EDuration="1.975415202s" podCreationTimestamp="2025-10-04 04:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:45.928671101 +0000 UTC m=+5677.220680813" watchObservedRunningTime="2025-10-04 04:37:45.975415202 +0000 UTC m=+5677.267424924" Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.003101 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.003083856 podStartE2EDuration="2.003083856s" podCreationTimestamp="2025-10-04 04:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:45.950528711 +0000 UTC m=+5677.242538423" watchObservedRunningTime="2025-10-04 04:37:46.003083856 +0000 UTC m=+5677.295093568" Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.494364 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdzqf"] Oct 04 04:37:46 crc kubenswrapper[4770]: W1004 04:37:46.506703 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c093be9_71c5_4478_84b2_da4d87396fc9.slice/crio-5926effa75dc240ecfc0b5d82b8b5b683d2e47c696a37ffb21ff9c1530a9eb36 WatchSource:0}: Error finding container 5926effa75dc240ecfc0b5d82b8b5b683d2e47c696a37ffb21ff9c1530a9eb36: Status 404 returned error can't find the container with id 5926effa75dc240ecfc0b5d82b8b5b683d2e47c696a37ffb21ff9c1530a9eb36 Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.895268 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f51bb4-55de-41b6-a91a-9b7b965ac20c","Type":"ContainerStarted","Data":"8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae"} Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.901137 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" event={"ID":"facf4f26-1f71-4f73-9d66-aaeb7476628d","Type":"ContainerStarted","Data":"820c7c61b51c718bcbe2424572a15765e125f7805069efba11a7265a35577e82"} Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.901271 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.903568 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xdzqf" event={"ID":"3c093be9-71c5-4478-84b2-da4d87396fc9","Type":"ContainerStarted","Data":"1506ae294a3ccc8964a0cccf021ed83a748c273902116cc940de35ba2f635bd2"} Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.903602 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xdzqf" event={"ID":"3c093be9-71c5-4478-84b2-da4d87396fc9","Type":"ContainerStarted","Data":"5926effa75dc240ecfc0b5d82b8b5b683d2e47c696a37ffb21ff9c1530a9eb36"} Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.919963 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.919944089 podStartE2EDuration="2.919944089s" podCreationTimestamp="2025-10-04 04:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:46.917627518 +0000 UTC m=+5678.209637240" watchObservedRunningTime="2025-10-04 04:37:46.919944089 +0000 UTC m=+5678.211953821" Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.945587 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" podStartSLOduration=2.945571219 podStartE2EDuration="2.945571219s" podCreationTimestamp="2025-10-04 04:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:46.937378334 +0000 UTC m=+5678.229388046" watchObservedRunningTime="2025-10-04 04:37:46.945571219 +0000 UTC m=+5678.237580931" Oct 04 04:37:46 crc kubenswrapper[4770]: I1004 04:37:46.963637 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-xdzqf" podStartSLOduration=1.96362105 podStartE2EDuration="1.96362105s" podCreationTimestamp="2025-10-04 04:37:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:46.959464622 +0000 UTC m=+5678.251474334" watchObservedRunningTime="2025-10-04 04:37:46.96362105 +0000 UTC m=+5678.255630762" Oct 04 04:37:49 crc kubenswrapper[4770]: I1004 04:37:49.557632 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:37:49 crc kubenswrapper[4770]: I1004 04:37:49.659434 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:37:49 crc kubenswrapper[4770]: I1004 04:37:49.659790 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:37:49 crc kubenswrapper[4770]: I1004 04:37:49.730594 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:49 crc kubenswrapper[4770]: I1004 04:37:49.933565 4770 generic.go:334] "Generic (PLEG): container finished" podID="3c093be9-71c5-4478-84b2-da4d87396fc9" containerID="1506ae294a3ccc8964a0cccf021ed83a748c273902116cc940de35ba2f635bd2" exitCode=0 Oct 04 04:37:49 crc kubenswrapper[4770]: I1004 04:37:49.934619 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xdzqf" event={"ID":"3c093be9-71c5-4478-84b2-da4d87396fc9","Type":"ContainerDied","Data":"1506ae294a3ccc8964a0cccf021ed83a748c273902116cc940de35ba2f635bd2"} Oct 04 04:37:50 crc kubenswrapper[4770]: I1004 04:37:50.943225 4770 generic.go:334] "Generic (PLEG): container finished" podID="38e50e99-f55d-47a6-ab20-fc4c37c91852" containerID="abb7af7395e58bb4795f53134cc7c15f7698cbe539b0c93e346617e09db55e5c" exitCode=0 Oct 04 04:37:50 crc kubenswrapper[4770]: I1004 04:37:50.943319 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gpw5q" event={"ID":"38e50e99-f55d-47a6-ab20-fc4c37c91852","Type":"ContainerDied","Data":"abb7af7395e58bb4795f53134cc7c15f7698cbe539b0c93e346617e09db55e5c"} Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.349671 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.424286 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-combined-ca-bundle\") pod \"3c093be9-71c5-4478-84b2-da4d87396fc9\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.424374 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-config-data\") pod \"3c093be9-71c5-4478-84b2-da4d87396fc9\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.424422 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-scripts\") pod \"3c093be9-71c5-4478-84b2-da4d87396fc9\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.424485 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b25r\" (UniqueName: \"kubernetes.io/projected/3c093be9-71c5-4478-84b2-da4d87396fc9-kube-api-access-6b25r\") pod \"3c093be9-71c5-4478-84b2-da4d87396fc9\" (UID: \"3c093be9-71c5-4478-84b2-da4d87396fc9\") " Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.434299 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-scripts" (OuterVolumeSpecName: "scripts") pod "3c093be9-71c5-4478-84b2-da4d87396fc9" (UID: "3c093be9-71c5-4478-84b2-da4d87396fc9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.434394 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c093be9-71c5-4478-84b2-da4d87396fc9-kube-api-access-6b25r" (OuterVolumeSpecName: "kube-api-access-6b25r") pod "3c093be9-71c5-4478-84b2-da4d87396fc9" (UID: "3c093be9-71c5-4478-84b2-da4d87396fc9"). InnerVolumeSpecName "kube-api-access-6b25r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.454363 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-config-data" (OuterVolumeSpecName: "config-data") pod "3c093be9-71c5-4478-84b2-da4d87396fc9" (UID: "3c093be9-71c5-4478-84b2-da4d87396fc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.454897 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c093be9-71c5-4478-84b2-da4d87396fc9" (UID: "3c093be9-71c5-4478-84b2-da4d87396fc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.526938 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.526971 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.526981 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c093be9-71c5-4478-84b2-da4d87396fc9-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.526991 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b25r\" (UniqueName: \"kubernetes.io/projected/3c093be9-71c5-4478-84b2-da4d87396fc9-kube-api-access-6b25r\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.955391 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xdzqf" Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.955575 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xdzqf" event={"ID":"3c093be9-71c5-4478-84b2-da4d87396fc9","Type":"ContainerDied","Data":"5926effa75dc240ecfc0b5d82b8b5b683d2e47c696a37ffb21ff9c1530a9eb36"} Oct 04 04:37:51 crc kubenswrapper[4770]: I1004 04:37:51.955761 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5926effa75dc240ecfc0b5d82b8b5b683d2e47c696a37ffb21ff9c1530a9eb36" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.027283 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:37:52 crc kubenswrapper[4770]: E1004 04:37:52.027773 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c093be9-71c5-4478-84b2-da4d87396fc9" containerName="nova-cell1-conductor-db-sync" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.027828 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c093be9-71c5-4478-84b2-da4d87396fc9" containerName="nova-cell1-conductor-db-sync" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.028032 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c093be9-71c5-4478-84b2-da4d87396fc9" containerName="nova-cell1-conductor-db-sync" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.028753 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.039403 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.042350 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.137990 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.138050 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.138313 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv5hn\" (UniqueName: \"kubernetes.io/projected/50830e0d-eeca-4ee8-ad7a-adca578cb88f-kube-api-access-qv5hn\") pod \"nova-cell1-conductor-0\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.239743 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.239837 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv5hn\" (UniqueName: \"kubernetes.io/projected/50830e0d-eeca-4ee8-ad7a-adca578cb88f-kube-api-access-qv5hn\") pod \"nova-cell1-conductor-0\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.239931 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.243767 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.261458 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv5hn\" (UniqueName: \"kubernetes.io/projected/50830e0d-eeca-4ee8-ad7a-adca578cb88f-kube-api-access-qv5hn\") pod \"nova-cell1-conductor-0\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.262790 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.346681 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.349889 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.442408 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-scripts\") pod \"38e50e99-f55d-47a6-ab20-fc4c37c91852\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.442855 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-config-data\") pod \"38e50e99-f55d-47a6-ab20-fc4c37c91852\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.442972 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-combined-ca-bundle\") pod \"38e50e99-f55d-47a6-ab20-fc4c37c91852\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.443079 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thwnw\" (UniqueName: \"kubernetes.io/projected/38e50e99-f55d-47a6-ab20-fc4c37c91852-kube-api-access-thwnw\") pod \"38e50e99-f55d-47a6-ab20-fc4c37c91852\" (UID: \"38e50e99-f55d-47a6-ab20-fc4c37c91852\") " Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.449819 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-scripts" (OuterVolumeSpecName: "scripts") pod "38e50e99-f55d-47a6-ab20-fc4c37c91852" (UID: "38e50e99-f55d-47a6-ab20-fc4c37c91852"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.450183 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38e50e99-f55d-47a6-ab20-fc4c37c91852-kube-api-access-thwnw" (OuterVolumeSpecName: "kube-api-access-thwnw") pod "38e50e99-f55d-47a6-ab20-fc4c37c91852" (UID: "38e50e99-f55d-47a6-ab20-fc4c37c91852"). InnerVolumeSpecName "kube-api-access-thwnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.472579 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38e50e99-f55d-47a6-ab20-fc4c37c91852" (UID: "38e50e99-f55d-47a6-ab20-fc4c37c91852"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.477186 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-config-data" (OuterVolumeSpecName: "config-data") pod "38e50e99-f55d-47a6-ab20-fc4c37c91852" (UID: "38e50e99-f55d-47a6-ab20-fc4c37c91852"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.546268 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.546543 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.546562 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thwnw\" (UniqueName: \"kubernetes.io/projected/38e50e99-f55d-47a6-ab20-fc4c37c91852-kube-api-access-thwnw\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.546598 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38e50e99-f55d-47a6-ab20-fc4c37c91852-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.787130 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:37:52 crc kubenswrapper[4770]: W1004 04:37:52.790972 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50830e0d_eeca_4ee8_ad7a_adca578cb88f.slice/crio-c4f8165225bcb96b751a096646c2c6025942311f56f5385d05d3653376e3b325 WatchSource:0}: Error finding container c4f8165225bcb96b751a096646c2c6025942311f56f5385d05d3653376e3b325: Status 404 returned error can't find the container with id c4f8165225bcb96b751a096646c2c6025942311f56f5385d05d3653376e3b325 Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.968721 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"50830e0d-eeca-4ee8-ad7a-adca578cb88f","Type":"ContainerStarted","Data":"c4f8165225bcb96b751a096646c2c6025942311f56f5385d05d3653376e3b325"} Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.971143 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gpw5q" event={"ID":"38e50e99-f55d-47a6-ab20-fc4c37c91852","Type":"ContainerDied","Data":"007217b67c6ff7295f0c24ecedb69e0d8575642a23b04c8518823e0038df8e60"} Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.971176 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="007217b67c6ff7295f0c24ecedb69e0d8575642a23b04c8518823e0038df8e60" Oct 04 04:37:52 crc kubenswrapper[4770]: I1004 04:37:52.971246 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gpw5q" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.157901 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.158349 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerName="nova-api-log" containerID="cri-o://bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608" gracePeriod=30 Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.158424 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerName="nova-api-api" containerID="cri-o://3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea" gracePeriod=30 Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.175246 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.175562 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="499c7c37-0517-4c0d-94cd-f0af00466cef" containerName="nova-scheduler-scheduler" containerID="cri-o://1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea" gracePeriod=30 Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.186407 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.186838 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerName="nova-metadata-metadata" containerID="cri-o://8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae" gracePeriod=30 Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.186715 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerName="nova-metadata-log" containerID="cri-o://7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990" gracePeriod=30 Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.830615 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.836640 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867242 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-config-data\") pod \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867311 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9xxw\" (UniqueName: \"kubernetes.io/projected/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-kube-api-access-w9xxw\") pod \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867344 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-config-data\") pod \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867376 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-combined-ca-bundle\") pod \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867429 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f51bb4-55de-41b6-a91a-9b7b965ac20c-logs\") pod \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867485 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-logs\") pod \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\" (UID: \"ecc5739f-51fe-4aee-a31f-2c8654fd9edf\") " Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867517 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-combined-ca-bundle\") pod \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867565 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v6g5\" (UniqueName: \"kubernetes.io/projected/17f51bb4-55de-41b6-a91a-9b7b965ac20c-kube-api-access-9v6g5\") pod \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\" (UID: \"17f51bb4-55de-41b6-a91a-9b7b965ac20c\") " Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867804 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17f51bb4-55de-41b6-a91a-9b7b965ac20c-logs" (OuterVolumeSpecName: "logs") pod "17f51bb4-55de-41b6-a91a-9b7b965ac20c" (UID: "17f51bb4-55de-41b6-a91a-9b7b965ac20c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.867814 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-logs" (OuterVolumeSpecName: "logs") pod "ecc5739f-51fe-4aee-a31f-2c8654fd9edf" (UID: "ecc5739f-51fe-4aee-a31f-2c8654fd9edf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.868127 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.868153 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f51bb4-55de-41b6-a91a-9b7b965ac20c-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.874072 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-kube-api-access-w9xxw" (OuterVolumeSpecName: "kube-api-access-w9xxw") pod "ecc5739f-51fe-4aee-a31f-2c8654fd9edf" (UID: "ecc5739f-51fe-4aee-a31f-2c8654fd9edf"). InnerVolumeSpecName "kube-api-access-w9xxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.875121 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f51bb4-55de-41b6-a91a-9b7b965ac20c-kube-api-access-9v6g5" (OuterVolumeSpecName: "kube-api-access-9v6g5") pod "17f51bb4-55de-41b6-a91a-9b7b965ac20c" (UID: "17f51bb4-55de-41b6-a91a-9b7b965ac20c"). InnerVolumeSpecName "kube-api-access-9v6g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.898448 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17f51bb4-55de-41b6-a91a-9b7b965ac20c" (UID: "17f51bb4-55de-41b6-a91a-9b7b965ac20c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.899272 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-config-data" (OuterVolumeSpecName: "config-data") pod "17f51bb4-55de-41b6-a91a-9b7b965ac20c" (UID: "17f51bb4-55de-41b6-a91a-9b7b965ac20c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.899361 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-config-data" (OuterVolumeSpecName: "config-data") pod "ecc5739f-51fe-4aee-a31f-2c8654fd9edf" (UID: "ecc5739f-51fe-4aee-a31f-2c8654fd9edf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.902462 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecc5739f-51fe-4aee-a31f-2c8654fd9edf" (UID: "ecc5739f-51fe-4aee-a31f-2c8654fd9edf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.970257 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.970290 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.970300 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v6g5\" (UniqueName: \"kubernetes.io/projected/17f51bb4-55de-41b6-a91a-9b7b965ac20c-kube-api-access-9v6g5\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.970310 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.970318 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9xxw\" (UniqueName: \"kubernetes.io/projected/ecc5739f-51fe-4aee-a31f-2c8654fd9edf-kube-api-access-w9xxw\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.970325 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f51bb4-55de-41b6-a91a-9b7b965ac20c-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.982416 4770 generic.go:334] "Generic (PLEG): container finished" podID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerID="3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea" exitCode=0 Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.982453 4770 generic.go:334] "Generic (PLEG): container finished" podID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerID="bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608" exitCode=143 Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.982484 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.982547 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ecc5739f-51fe-4aee-a31f-2c8654fd9edf","Type":"ContainerDied","Data":"3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea"} Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.982627 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ecc5739f-51fe-4aee-a31f-2c8654fd9edf","Type":"ContainerDied","Data":"bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608"} Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.982640 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ecc5739f-51fe-4aee-a31f-2c8654fd9edf","Type":"ContainerDied","Data":"7492a9a1a4be09a52546603d2bea783742b42b146430da37bfcac7e70eed1bcf"} Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.982678 4770 scope.go:117] "RemoveContainer" containerID="3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.984735 4770 generic.go:334] "Generic (PLEG): container finished" podID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerID="8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae" exitCode=0 Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.984767 4770 generic.go:334] "Generic (PLEG): container finished" podID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerID="7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990" exitCode=143 Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.984826 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f51bb4-55de-41b6-a91a-9b7b965ac20c","Type":"ContainerDied","Data":"8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae"} Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.985649 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f51bb4-55de-41b6-a91a-9b7b965ac20c","Type":"ContainerDied","Data":"7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990"} Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.985675 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f51bb4-55de-41b6-a91a-9b7b965ac20c","Type":"ContainerDied","Data":"385cb8d729bc08b6ffccc9beaf380a8413f56a40701225f9a87346f280fdde1f"} Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.984915 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.989346 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"50830e0d-eeca-4ee8-ad7a-adca578cb88f","Type":"ContainerStarted","Data":"20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b"} Oct 04 04:37:53 crc kubenswrapper[4770]: I1004 04:37:53.990205 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.012268 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.012244055 podStartE2EDuration="2.012244055s" podCreationTimestamp="2025-10-04 04:37:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:54.006453243 +0000 UTC m=+5685.298462955" watchObservedRunningTime="2025-10-04 04:37:54.012244055 +0000 UTC m=+5685.304253767" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.015219 4770 scope.go:117] "RemoveContainer" containerID="bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.051157 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.059758 4770 scope.go:117] "RemoveContainer" containerID="3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.061950 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: E1004 04:37:54.062058 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea\": container with ID starting with 3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea not found: ID does not exist" containerID="3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.062095 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea"} err="failed to get container status \"3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea\": rpc error: code = NotFound desc = could not find container \"3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea\": container with ID starting with 3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea not found: ID does not exist" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.062140 4770 scope.go:117] "RemoveContainer" containerID="bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608" Oct 04 04:37:54 crc kubenswrapper[4770]: E1004 04:37:54.062405 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608\": container with ID starting with bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608 not found: ID does not exist" containerID="bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.062435 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608"} err="failed to get container status \"bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608\": rpc error: code = NotFound desc = could not find container \"bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608\": container with ID starting with bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608 not found: ID does not exist" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.062456 4770 scope.go:117] "RemoveContainer" containerID="3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.063060 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea"} err="failed to get container status \"3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea\": rpc error: code = NotFound desc = could not find container \"3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea\": container with ID starting with 3c9275a2b9a12d70d0f552d4eb83390f4128c24c7bb5620c790872e48dba13ea not found: ID does not exist" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.063079 4770 scope.go:117] "RemoveContainer" containerID="bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.063275 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608"} err="failed to get container status \"bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608\": rpc error: code = NotFound desc = could not find container \"bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608\": container with ID starting with bcbda165a62a7e42d336f720227959102bf687aa495638b7921f87924ec10608 not found: ID does not exist" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.063312 4770 scope.go:117] "RemoveContainer" containerID="8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.068857 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.083168 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105065 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: E1004 04:37:54.105634 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerName="nova-api-api" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105664 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerName="nova-api-api" Oct 04 04:37:54 crc kubenswrapper[4770]: E1004 04:37:54.105680 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerName="nova-metadata-metadata" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105687 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerName="nova-metadata-metadata" Oct 04 04:37:54 crc kubenswrapper[4770]: E1004 04:37:54.105696 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerName="nova-metadata-log" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105701 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerName="nova-metadata-log" Oct 04 04:37:54 crc kubenswrapper[4770]: E1004 04:37:54.105722 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerName="nova-api-log" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105727 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerName="nova-api-log" Oct 04 04:37:54 crc kubenswrapper[4770]: E1004 04:37:54.105742 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38e50e99-f55d-47a6-ab20-fc4c37c91852" containerName="nova-manage" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105748 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="38e50e99-f55d-47a6-ab20-fc4c37c91852" containerName="nova-manage" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105930 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerName="nova-api-api" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105944 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="38e50e99-f55d-47a6-ab20-fc4c37c91852" containerName="nova-manage" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105954 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerName="nova-metadata-log" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105964 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" containerName="nova-api-log" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.105987 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" containerName="nova-metadata-metadata" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.106883 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.110407 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.112521 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.114780 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.117181 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.118611 4770 scope.go:117] "RemoveContainer" containerID="7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.128615 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.138539 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.152499 4770 scope.go:117] "RemoveContainer" containerID="8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae" Oct 04 04:37:54 crc kubenswrapper[4770]: E1004 04:37:54.152974 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae\": container with ID starting with 8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae not found: ID does not exist" containerID="8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.153024 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae"} err="failed to get container status \"8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae\": rpc error: code = NotFound desc = could not find container \"8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae\": container with ID starting with 8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae not found: ID does not exist" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.153052 4770 scope.go:117] "RemoveContainer" containerID="7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990" Oct 04 04:37:54 crc kubenswrapper[4770]: E1004 04:37:54.153558 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990\": container with ID starting with 7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990 not found: ID does not exist" containerID="7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.153612 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990"} err="failed to get container status \"7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990\": rpc error: code = NotFound desc = could not find container \"7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990\": container with ID starting with 7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990 not found: ID does not exist" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.153638 4770 scope.go:117] "RemoveContainer" containerID="8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.154485 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae"} err="failed to get container status \"8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae\": rpc error: code = NotFound desc = could not find container \"8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae\": container with ID starting with 8367ae4e9d95dc8cbd0f3d6b2c98d0d50ab731f776d74e54e1fe7d7d30bb17ae not found: ID does not exist" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.154510 4770 scope.go:117] "RemoveContainer" containerID="7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.154744 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990"} err="failed to get container status \"7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990\": rpc error: code = NotFound desc = could not find container \"7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990\": container with ID starting with 7043255eb6697ebccac4bc494eadc23042e8da9e51d59f7420d820a18fde7990 not found: ID does not exist" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.182340 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-config-data\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.182418 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-config-data\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.182455 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.182480 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcmbp\" (UniqueName: \"kubernetes.io/projected/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-kube-api-access-hcmbp\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.182536 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.182561 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42x4z\" (UniqueName: \"kubernetes.io/projected/2ab59ae1-24dc-49f6-b109-f12204ba4e38-kube-api-access-42x4z\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.182602 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-logs\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.182782 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab59ae1-24dc-49f6-b109-f12204ba4e38-logs\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.283927 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab59ae1-24dc-49f6-b109-f12204ba4e38-logs\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.284019 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-config-data\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.284059 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-config-data\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.284086 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.284107 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcmbp\" (UniqueName: \"kubernetes.io/projected/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-kube-api-access-hcmbp\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.284140 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.284158 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42x4z\" (UniqueName: \"kubernetes.io/projected/2ab59ae1-24dc-49f6-b109-f12204ba4e38-kube-api-access-42x4z\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.284190 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-logs\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.284625 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-logs\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.284635 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab59ae1-24dc-49f6-b109-f12204ba4e38-logs\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.289528 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-config-data\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.289674 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.300673 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.300916 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-config-data\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.303551 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcmbp\" (UniqueName: \"kubernetes.io/projected/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-kube-api-access-hcmbp\") pod \"nova-metadata-0\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.303702 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42x4z\" (UniqueName: \"kubernetes.io/projected/2ab59ae1-24dc-49f6-b109-f12204ba4e38-kube-api-access-42x4z\") pod \"nova-api-0\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.429116 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.444768 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.730496 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.747258 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.773258 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.860292 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85467b57bf-hjkz7"] Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.860867 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" podUID="455e4130-03c6-4e06-a9f5-4df56558327f" containerName="dnsmasq-dns" containerID="cri-o://3b435a70ab461749e918093686c1a3f680f394bb87307f2df6c03621b997870d" gracePeriod=10 Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.927302 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: W1004 04:37:54.957831 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60006b51_98cb_4d97_8f4a_fb71c4e2b5fb.slice/crio-a5dd92c0d288e5f838b3015aff9871aea71855980d167fdf0529d404a584d738 WatchSource:0}: Error finding container a5dd92c0d288e5f838b3015aff9871aea71855980d167fdf0529d404a584d738: Status 404 returned error can't find the container with id a5dd92c0d288e5f838b3015aff9871aea71855980d167fdf0529d404a584d738 Oct 04 04:37:54 crc kubenswrapper[4770]: I1004 04:37:54.980198 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:37:54 crc kubenswrapper[4770]: W1004 04:37:54.996994 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ab59ae1_24dc_49f6_b109_f12204ba4e38.slice/crio-0c671a24fc519c5262ee91686d16ada663fa88582485506ceeaadb13041e2bd3 WatchSource:0}: Error finding container 0c671a24fc519c5262ee91686d16ada663fa88582485506ceeaadb13041e2bd3: Status 404 returned error can't find the container with id 0c671a24fc519c5262ee91686d16ada663fa88582485506ceeaadb13041e2bd3 Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.002287 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb","Type":"ContainerStarted","Data":"a5dd92c0d288e5f838b3015aff9871aea71855980d167fdf0529d404a584d738"} Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.005219 4770 generic.go:334] "Generic (PLEG): container finished" podID="455e4130-03c6-4e06-a9f5-4df56558327f" containerID="3b435a70ab461749e918093686c1a3f680f394bb87307f2df6c03621b997870d" exitCode=0 Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.006089 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" event={"ID":"455e4130-03c6-4e06-a9f5-4df56558327f","Type":"ContainerDied","Data":"3b435a70ab461749e918093686c1a3f680f394bb87307f2df6c03621b997870d"} Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.015545 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.444233 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.514751 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwjgp\" (UniqueName: \"kubernetes.io/projected/455e4130-03c6-4e06-a9f5-4df56558327f-kube-api-access-nwjgp\") pod \"455e4130-03c6-4e06-a9f5-4df56558327f\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.514845 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-sb\") pod \"455e4130-03c6-4e06-a9f5-4df56558327f\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.514893 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-nb\") pod \"455e4130-03c6-4e06-a9f5-4df56558327f\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.514984 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-dns-svc\") pod \"455e4130-03c6-4e06-a9f5-4df56558327f\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.515093 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-config\") pod \"455e4130-03c6-4e06-a9f5-4df56558327f\" (UID: \"455e4130-03c6-4e06-a9f5-4df56558327f\") " Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.521323 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/455e4130-03c6-4e06-a9f5-4df56558327f-kube-api-access-nwjgp" (OuterVolumeSpecName: "kube-api-access-nwjgp") pod "455e4130-03c6-4e06-a9f5-4df56558327f" (UID: "455e4130-03c6-4e06-a9f5-4df56558327f"). InnerVolumeSpecName "kube-api-access-nwjgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.564860 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-config" (OuterVolumeSpecName: "config") pod "455e4130-03c6-4e06-a9f5-4df56558327f" (UID: "455e4130-03c6-4e06-a9f5-4df56558327f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.569083 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "455e4130-03c6-4e06-a9f5-4df56558327f" (UID: "455e4130-03c6-4e06-a9f5-4df56558327f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.583649 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "455e4130-03c6-4e06-a9f5-4df56558327f" (UID: "455e4130-03c6-4e06-a9f5-4df56558327f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.590233 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "455e4130-03c6-4e06-a9f5-4df56558327f" (UID: "455e4130-03c6-4e06-a9f5-4df56558327f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.624929 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.624974 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.624987 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwjgp\" (UniqueName: \"kubernetes.io/projected/455e4130-03c6-4e06-a9f5-4df56558327f-kube-api-access-nwjgp\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.624997 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.625019 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/455e4130-03c6-4e06-a9f5-4df56558327f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.683734 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17f51bb4-55de-41b6-a91a-9b7b965ac20c" path="/var/lib/kubelet/pods/17f51bb4-55de-41b6-a91a-9b7b965ac20c/volumes" Oct 04 04:37:55 crc kubenswrapper[4770]: I1004 04:37:55.684341 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecc5739f-51fe-4aee-a31f-2c8654fd9edf" path="/var/lib/kubelet/pods/ecc5739f-51fe-4aee-a31f-2c8654fd9edf/volumes" Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.020163 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ab59ae1-24dc-49f6-b109-f12204ba4e38","Type":"ContainerStarted","Data":"e26bd6e9663e524661e1621223d79bed889d994cc4a00d41e73e0ccaa6ea5745"} Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.020539 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ab59ae1-24dc-49f6-b109-f12204ba4e38","Type":"ContainerStarted","Data":"1012b28f28d58fe59a0639ac7be5b695b2240b3e28ff5e76dbd89b9c2c3c812b"} Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.020556 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ab59ae1-24dc-49f6-b109-f12204ba4e38","Type":"ContainerStarted","Data":"0c671a24fc519c5262ee91686d16ada663fa88582485506ceeaadb13041e2bd3"} Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.022725 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb","Type":"ContainerStarted","Data":"ede9786b26fd9af96e9899c0d6294943617452f1e5ea820df101c54ca1a1e198"} Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.022773 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb","Type":"ContainerStarted","Data":"72a4e783e08af8f179860f61d0026cebedbf48ebe9d9ec2186f5e449ab6635ab"} Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.024915 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.024939 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85467b57bf-hjkz7" event={"ID":"455e4130-03c6-4e06-a9f5-4df56558327f","Type":"ContainerDied","Data":"b9f2a9209d86c861b178c1e6e31e60cc2c19d7b6221eba5ef9d87029d0af9eea"} Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.024972 4770 scope.go:117] "RemoveContainer" containerID="3b435a70ab461749e918093686c1a3f680f394bb87307f2df6c03621b997870d" Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.046442 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.04641468 podStartE2EDuration="2.04641468s" podCreationTimestamp="2025-10-04 04:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:56.037355383 +0000 UTC m=+5687.329365095" watchObservedRunningTime="2025-10-04 04:37:56.04641468 +0000 UTC m=+5687.338424392" Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.054072 4770 scope.go:117] "RemoveContainer" containerID="606e187a5ce7dba44d4054c6ecc4c8597e9c2eadeedde24e024b4cbb0fe7fd7a" Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.066103 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.066083274 podStartE2EDuration="2.066083274s" podCreationTimestamp="2025-10-04 04:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:56.058538307 +0000 UTC m=+5687.350548019" watchObservedRunningTime="2025-10-04 04:37:56.066083274 +0000 UTC m=+5687.358092986" Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.079669 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85467b57bf-hjkz7"] Oct 04 04:37:56 crc kubenswrapper[4770]: I1004 04:37:56.085918 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85467b57bf-hjkz7"] Oct 04 04:37:57 crc kubenswrapper[4770]: I1004 04:37:57.690658 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="455e4130-03c6-4e06-a9f5-4df56558327f" path="/var/lib/kubelet/pods/455e4130-03c6-4e06-a9f5-4df56558327f/volumes" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.009886 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.097259 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-combined-ca-bundle\") pod \"499c7c37-0517-4c0d-94cd-f0af00466cef\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.097308 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-config-data\") pod \"499c7c37-0517-4c0d-94cd-f0af00466cef\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.097349 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqzgw\" (UniqueName: \"kubernetes.io/projected/499c7c37-0517-4c0d-94cd-f0af00466cef-kube-api-access-bqzgw\") pod \"499c7c37-0517-4c0d-94cd-f0af00466cef\" (UID: \"499c7c37-0517-4c0d-94cd-f0af00466cef\") " Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.102065 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/499c7c37-0517-4c0d-94cd-f0af00466cef-kube-api-access-bqzgw" (OuterVolumeSpecName: "kube-api-access-bqzgw") pod "499c7c37-0517-4c0d-94cd-f0af00466cef" (UID: "499c7c37-0517-4c0d-94cd-f0af00466cef"). InnerVolumeSpecName "kube-api-access-bqzgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.103474 4770 generic.go:334] "Generic (PLEG): container finished" podID="499c7c37-0517-4c0d-94cd-f0af00466cef" containerID="1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea" exitCode=0 Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.103529 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"499c7c37-0517-4c0d-94cd-f0af00466cef","Type":"ContainerDied","Data":"1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea"} Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.103554 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"499c7c37-0517-4c0d-94cd-f0af00466cef","Type":"ContainerDied","Data":"cf27ea9914999e75a81ff2b1a04d9c0c503692018964ff91a9947aabee98763e"} Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.103592 4770 scope.go:117] "RemoveContainer" containerID="1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.103740 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.121374 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-config-data" (OuterVolumeSpecName: "config-data") pod "499c7c37-0517-4c0d-94cd-f0af00466cef" (UID: "499c7c37-0517-4c0d-94cd-f0af00466cef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.136248 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "499c7c37-0517-4c0d-94cd-f0af00466cef" (UID: "499c7c37-0517-4c0d-94cd-f0af00466cef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.204532 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.204874 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/499c7c37-0517-4c0d-94cd-f0af00466cef-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.204900 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqzgw\" (UniqueName: \"kubernetes.io/projected/499c7c37-0517-4c0d-94cd-f0af00466cef-kube-api-access-bqzgw\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.205947 4770 scope.go:117] "RemoveContainer" containerID="1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea" Oct 04 04:37:58 crc kubenswrapper[4770]: E1004 04:37:58.206503 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea\": container with ID starting with 1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea not found: ID does not exist" containerID="1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.206543 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea"} err="failed to get container status \"1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea\": rpc error: code = NotFound desc = could not find container \"1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea\": container with ID starting with 1ecfa6d437e882a4158c9521a46530086fe6061f6edad33350ce2712e43671ea not found: ID does not exist" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.446073 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.456514 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.464031 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:37:58 crc kubenswrapper[4770]: E1004 04:37:58.464439 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="455e4130-03c6-4e06-a9f5-4df56558327f" containerName="init" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.464460 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="455e4130-03c6-4e06-a9f5-4df56558327f" containerName="init" Oct 04 04:37:58 crc kubenswrapper[4770]: E1004 04:37:58.464480 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="499c7c37-0517-4c0d-94cd-f0af00466cef" containerName="nova-scheduler-scheduler" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.464487 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="499c7c37-0517-4c0d-94cd-f0af00466cef" containerName="nova-scheduler-scheduler" Oct 04 04:37:58 crc kubenswrapper[4770]: E1004 04:37:58.464502 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="455e4130-03c6-4e06-a9f5-4df56558327f" containerName="dnsmasq-dns" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.464508 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="455e4130-03c6-4e06-a9f5-4df56558327f" containerName="dnsmasq-dns" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.464675 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="499c7c37-0517-4c0d-94cd-f0af00466cef" containerName="nova-scheduler-scheduler" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.464689 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="455e4130-03c6-4e06-a9f5-4df56558327f" containerName="dnsmasq-dns" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.465450 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.468661 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.471207 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.611609 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-config-data\") pod \"nova-scheduler-0\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.611679 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7l62\" (UniqueName: \"kubernetes.io/projected/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-kube-api-access-g7l62\") pod \"nova-scheduler-0\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.611723 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.712778 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7l62\" (UniqueName: \"kubernetes.io/projected/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-kube-api-access-g7l62\") pod \"nova-scheduler-0\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.712845 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.712928 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-config-data\") pod \"nova-scheduler-0\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.716917 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.718120 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-config-data\") pod \"nova-scheduler-0\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.736653 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7l62\" (UniqueName: \"kubernetes.io/projected/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-kube-api-access-g7l62\") pod \"nova-scheduler-0\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " pod="openstack/nova-scheduler-0" Oct 04 04:37:58 crc kubenswrapper[4770]: I1004 04:37:58.790704 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:37:59 crc kubenswrapper[4770]: I1004 04:37:59.312370 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:37:59 crc kubenswrapper[4770]: I1004 04:37:59.430549 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:37:59 crc kubenswrapper[4770]: I1004 04:37:59.430654 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:37:59 crc kubenswrapper[4770]: I1004 04:37:59.687683 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="499c7c37-0517-4c0d-94cd-f0af00466cef" path="/var/lib/kubelet/pods/499c7c37-0517-4c0d-94cd-f0af00466cef/volumes" Oct 04 04:38:00 crc kubenswrapper[4770]: I1004 04:38:00.123982 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16","Type":"ContainerStarted","Data":"f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e"} Oct 04 04:38:00 crc kubenswrapper[4770]: I1004 04:38:00.124277 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16","Type":"ContainerStarted","Data":"437c8b8fc751dd1bf5881ac859a8f432ddcc5a2ef118da43eb8f1160b3190134"} Oct 04 04:38:00 crc kubenswrapper[4770]: I1004 04:38:00.155237 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.155212947 podStartE2EDuration="2.155212947s" podCreationTimestamp="2025-10-04 04:37:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:38:00.145330599 +0000 UTC m=+5691.437340341" watchObservedRunningTime="2025-10-04 04:38:00.155212947 +0000 UTC m=+5691.447222659" Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.384792 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.925989 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-c6n99"] Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.927488 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.929947 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.929947 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.992549 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-c6n99"] Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.995936 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.995989 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-config-data\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.996297 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-scripts\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:02 crc kubenswrapper[4770]: I1004 04:38:02.996482 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdjx4\" (UniqueName: \"kubernetes.io/projected/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-kube-api-access-kdjx4\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.098217 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.099119 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-config-data\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.099209 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-scripts\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.099288 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdjx4\" (UniqueName: \"kubernetes.io/projected/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-kube-api-access-kdjx4\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.103286 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-scripts\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.103446 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-config-data\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.105068 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.117982 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdjx4\" (UniqueName: \"kubernetes.io/projected/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-kube-api-access-kdjx4\") pod \"nova-cell1-cell-mapping-c6n99\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.297761 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.790973 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:38:03 crc kubenswrapper[4770]: I1004 04:38:03.927947 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-c6n99"] Oct 04 04:38:04 crc kubenswrapper[4770]: I1004 04:38:04.167481 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c6n99" event={"ID":"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819","Type":"ContainerStarted","Data":"b8462b174804162f97a91e07e3708a79afb7c7c4f0cc931b0a50061aa5d74877"} Oct 04 04:38:04 crc kubenswrapper[4770]: I1004 04:38:04.430728 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:38:04 crc kubenswrapper[4770]: I1004 04:38:04.431900 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:38:04 crc kubenswrapper[4770]: I1004 04:38:04.445778 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:38:04 crc kubenswrapper[4770]: I1004 04:38:04.445864 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:38:05 crc kubenswrapper[4770]: I1004 04:38:05.175599 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c6n99" event={"ID":"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819","Type":"ContainerStarted","Data":"7ecc1d7c90290f90a4ef3e75d7a375fce97a408ebc53a7ea3a700a530d8d1ff5"} Oct 04 04:38:05 crc kubenswrapper[4770]: I1004 04:38:05.196167 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-c6n99" podStartSLOduration=3.196149228 podStartE2EDuration="3.196149228s" podCreationTimestamp="2025-10-04 04:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:38:05.19013719 +0000 UTC m=+5696.482146932" watchObservedRunningTime="2025-10-04 04:38:05.196149228 +0000 UTC m=+5696.488158940" Oct 04 04:38:05 crc kubenswrapper[4770]: I1004 04:38:05.595201 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.68:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:38:05 crc kubenswrapper[4770]: I1004 04:38:05.595245 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.68:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:38:05 crc kubenswrapper[4770]: I1004 04:38:05.595269 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.69:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:38:05 crc kubenswrapper[4770]: I1004 04:38:05.595318 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.69:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:38:08 crc kubenswrapper[4770]: I1004 04:38:08.791659 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:38:08 crc kubenswrapper[4770]: I1004 04:38:08.818877 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:38:09 crc kubenswrapper[4770]: I1004 04:38:09.269452 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:38:10 crc kubenswrapper[4770]: I1004 04:38:10.245121 4770 generic.go:334] "Generic (PLEG): container finished" podID="34e7fc27-dfe7-4bb9-b6fc-b2fabc466819" containerID="7ecc1d7c90290f90a4ef3e75d7a375fce97a408ebc53a7ea3a700a530d8d1ff5" exitCode=0 Oct 04 04:38:10 crc kubenswrapper[4770]: I1004 04:38:10.246340 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c6n99" event={"ID":"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819","Type":"ContainerDied","Data":"7ecc1d7c90290f90a4ef3e75d7a375fce97a408ebc53a7ea3a700a530d8d1ff5"} Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.611479 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.658147 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-scripts\") pod \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.658275 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-config-data\") pod \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.658344 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-combined-ca-bundle\") pod \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.658389 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdjx4\" (UniqueName: \"kubernetes.io/projected/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-kube-api-access-kdjx4\") pod \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\" (UID: \"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819\") " Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.665471 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-kube-api-access-kdjx4" (OuterVolumeSpecName: "kube-api-access-kdjx4") pod "34e7fc27-dfe7-4bb9-b6fc-b2fabc466819" (UID: "34e7fc27-dfe7-4bb9-b6fc-b2fabc466819"). InnerVolumeSpecName "kube-api-access-kdjx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.665506 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-scripts" (OuterVolumeSpecName: "scripts") pod "34e7fc27-dfe7-4bb9-b6fc-b2fabc466819" (UID: "34e7fc27-dfe7-4bb9-b6fc-b2fabc466819"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.685973 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34e7fc27-dfe7-4bb9-b6fc-b2fabc466819" (UID: "34e7fc27-dfe7-4bb9-b6fc-b2fabc466819"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.700939 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-config-data" (OuterVolumeSpecName: "config-data") pod "34e7fc27-dfe7-4bb9-b6fc-b2fabc466819" (UID: "34e7fc27-dfe7-4bb9-b6fc-b2fabc466819"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.761142 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.761194 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.761212 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:11 crc kubenswrapper[4770]: I1004 04:38:11.761231 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdjx4\" (UniqueName: \"kubernetes.io/projected/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819-kube-api-access-kdjx4\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.262174 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c6n99" event={"ID":"34e7fc27-dfe7-4bb9-b6fc-b2fabc466819","Type":"ContainerDied","Data":"b8462b174804162f97a91e07e3708a79afb7c7c4f0cc931b0a50061aa5d74877"} Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.262226 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8462b174804162f97a91e07e3708a79afb7c7c4f0cc931b0a50061aa5d74877" Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.262236 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c6n99" Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.460138 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.460906 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-log" containerID="cri-o://1012b28f28d58fe59a0639ac7be5b695b2240b3e28ff5e76dbd89b9c2c3c812b" gracePeriod=30 Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.461046 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-api" containerID="cri-o://e26bd6e9663e524661e1621223d79bed889d994cc4a00d41e73e0ccaa6ea5745" gracePeriod=30 Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.480027 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.480292 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" containerName="nova-scheduler-scheduler" containerID="cri-o://f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e" gracePeriod=30 Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.494148 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.494432 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-log" containerID="cri-o://72a4e783e08af8f179860f61d0026cebedbf48ebe9d9ec2186f5e449ab6635ab" gracePeriod=30 Oct 04 04:38:12 crc kubenswrapper[4770]: I1004 04:38:12.494501 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-metadata" containerID="cri-o://ede9786b26fd9af96e9899c0d6294943617452f1e5ea820df101c54ca1a1e198" gracePeriod=30 Oct 04 04:38:13 crc kubenswrapper[4770]: I1004 04:38:13.274930 4770 generic.go:334] "Generic (PLEG): container finished" podID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerID="72a4e783e08af8f179860f61d0026cebedbf48ebe9d9ec2186f5e449ab6635ab" exitCode=143 Oct 04 04:38:13 crc kubenswrapper[4770]: I1004 04:38:13.275057 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb","Type":"ContainerDied","Data":"72a4e783e08af8f179860f61d0026cebedbf48ebe9d9ec2186f5e449ab6635ab"} Oct 04 04:38:13 crc kubenswrapper[4770]: I1004 04:38:13.277539 4770 generic.go:334] "Generic (PLEG): container finished" podID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerID="1012b28f28d58fe59a0639ac7be5b695b2240b3e28ff5e76dbd89b9c2c3c812b" exitCode=143 Oct 04 04:38:13 crc kubenswrapper[4770]: I1004 04:38:13.277593 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ab59ae1-24dc-49f6-b109-f12204ba4e38","Type":"ContainerDied","Data":"1012b28f28d58fe59a0639ac7be5b695b2240b3e28ff5e76dbd89b9c2c3c812b"} Oct 04 04:38:13 crc kubenswrapper[4770]: E1004 04:38:13.794430 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:38:13 crc kubenswrapper[4770]: E1004 04:38:13.796600 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:38:13 crc kubenswrapper[4770]: E1004 04:38:13.798162 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:38:13 crc kubenswrapper[4770]: E1004 04:38:13.798252 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" containerName="nova-scheduler-scheduler" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.308658 4770 generic.go:334] "Generic (PLEG): container finished" podID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerID="ede9786b26fd9af96e9899c0d6294943617452f1e5ea820df101c54ca1a1e198" exitCode=0 Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.309081 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb","Type":"ContainerDied","Data":"ede9786b26fd9af96e9899c0d6294943617452f1e5ea820df101c54ca1a1e198"} Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.311424 4770 generic.go:334] "Generic (PLEG): container finished" podID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerID="e26bd6e9663e524661e1621223d79bed889d994cc4a00d41e73e0ccaa6ea5745" exitCode=0 Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.311471 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ab59ae1-24dc-49f6-b109-f12204ba4e38","Type":"ContainerDied","Data":"e26bd6e9663e524661e1621223d79bed889d994cc4a00d41e73e0ccaa6ea5745"} Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.723857 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.731589 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.779528 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-config-data\") pod \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.780316 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-combined-ca-bundle\") pod \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.780398 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-logs\") pod \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.780457 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-config-data\") pod \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.780593 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-combined-ca-bundle\") pod \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.780636 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42x4z\" (UniqueName: \"kubernetes.io/projected/2ab59ae1-24dc-49f6-b109-f12204ba4e38-kube-api-access-42x4z\") pod \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.780657 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab59ae1-24dc-49f6-b109-f12204ba4e38-logs\") pod \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\" (UID: \"2ab59ae1-24dc-49f6-b109-f12204ba4e38\") " Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.780701 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcmbp\" (UniqueName: \"kubernetes.io/projected/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-kube-api-access-hcmbp\") pod \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\" (UID: \"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb\") " Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.781292 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ab59ae1-24dc-49f6-b109-f12204ba4e38-logs" (OuterVolumeSpecName: "logs") pod "2ab59ae1-24dc-49f6-b109-f12204ba4e38" (UID: "2ab59ae1-24dc-49f6-b109-f12204ba4e38"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.781540 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-logs" (OuterVolumeSpecName: "logs") pod "60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" (UID: "60006b51-98cb-4d97-8f4a-fb71c4e2b5fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.781664 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab59ae1-24dc-49f6-b109-f12204ba4e38-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.785905 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-kube-api-access-hcmbp" (OuterVolumeSpecName: "kube-api-access-hcmbp") pod "60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" (UID: "60006b51-98cb-4d97-8f4a-fb71c4e2b5fb"). InnerVolumeSpecName "kube-api-access-hcmbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.795387 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ab59ae1-24dc-49f6-b109-f12204ba4e38-kube-api-access-42x4z" (OuterVolumeSpecName: "kube-api-access-42x4z") pod "2ab59ae1-24dc-49f6-b109-f12204ba4e38" (UID: "2ab59ae1-24dc-49f6-b109-f12204ba4e38"). InnerVolumeSpecName "kube-api-access-42x4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.806166 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-config-data" (OuterVolumeSpecName: "config-data") pod "2ab59ae1-24dc-49f6-b109-f12204ba4e38" (UID: "2ab59ae1-24dc-49f6-b109-f12204ba4e38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.814311 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ab59ae1-24dc-49f6-b109-f12204ba4e38" (UID: "2ab59ae1-24dc-49f6-b109-f12204ba4e38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.821937 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-config-data" (OuterVolumeSpecName: "config-data") pod "60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" (UID: "60006b51-98cb-4d97-8f4a-fb71c4e2b5fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.834441 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" (UID: "60006b51-98cb-4d97-8f4a-fb71c4e2b5fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.883828 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.883869 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab59ae1-24dc-49f6-b109-f12204ba4e38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.883884 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42x4z\" (UniqueName: \"kubernetes.io/projected/2ab59ae1-24dc-49f6-b109-f12204ba4e38-kube-api-access-42x4z\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.883897 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcmbp\" (UniqueName: \"kubernetes.io/projected/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-kube-api-access-hcmbp\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.883908 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.883920 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4770]: I1004 04:38:16.883932 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.038730 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.087274 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-config-data\") pod \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.087325 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-combined-ca-bundle\") pod \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.087375 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7l62\" (UniqueName: \"kubernetes.io/projected/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-kube-api-access-g7l62\") pod \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\" (UID: \"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16\") " Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.096246 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-kube-api-access-g7l62" (OuterVolumeSpecName: "kube-api-access-g7l62") pod "58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" (UID: "58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16"). InnerVolumeSpecName "kube-api-access-g7l62". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.114976 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" (UID: "58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.140562 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-config-data" (OuterVolumeSpecName: "config-data") pod "58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" (UID: "58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.189837 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.189875 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.189889 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7l62\" (UniqueName: \"kubernetes.io/projected/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16-kube-api-access-g7l62\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.321111 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2ab59ae1-24dc-49f6-b109-f12204ba4e38","Type":"ContainerDied","Data":"0c671a24fc519c5262ee91686d16ada663fa88582485506ceeaadb13041e2bd3"} Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.321124 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.321183 4770 scope.go:117] "RemoveContainer" containerID="e26bd6e9663e524661e1621223d79bed889d994cc4a00d41e73e0ccaa6ea5745" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.322515 4770 generic.go:334] "Generic (PLEG): container finished" podID="58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" containerID="f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e" exitCode=0 Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.322584 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16","Type":"ContainerDied","Data":"f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e"} Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.322570 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.322605 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16","Type":"ContainerDied","Data":"437c8b8fc751dd1bf5881ac859a8f432ddcc5a2ef118da43eb8f1160b3190134"} Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.324676 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60006b51-98cb-4d97-8f4a-fb71c4e2b5fb","Type":"ContainerDied","Data":"a5dd92c0d288e5f838b3015aff9871aea71855980d167fdf0529d404a584d738"} Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.324789 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.343712 4770 scope.go:117] "RemoveContainer" containerID="1012b28f28d58fe59a0639ac7be5b695b2240b3e28ff5e76dbd89b9c2c3c812b" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.365524 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.381522 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.392322 4770 scope.go:117] "RemoveContainer" containerID="f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.393642 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.406282 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.428704 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: E1004 04:38:17.429138 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-log" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429157 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-log" Oct 04 04:38:17 crc kubenswrapper[4770]: E1004 04:38:17.429173 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-metadata" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429179 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-metadata" Oct 04 04:38:17 crc kubenswrapper[4770]: E1004 04:38:17.429196 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-api" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429203 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-api" Oct 04 04:38:17 crc kubenswrapper[4770]: E1004 04:38:17.429214 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e7fc27-dfe7-4bb9-b6fc-b2fabc466819" containerName="nova-manage" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429219 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e7fc27-dfe7-4bb9-b6fc-b2fabc466819" containerName="nova-manage" Oct 04 04:38:17 crc kubenswrapper[4770]: E1004 04:38:17.429232 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" containerName="nova-scheduler-scheduler" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429239 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" containerName="nova-scheduler-scheduler" Oct 04 04:38:17 crc kubenswrapper[4770]: E1004 04:38:17.429263 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-log" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429270 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-log" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429422 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-log" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429449 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="34e7fc27-dfe7-4bb9-b6fc-b2fabc466819" containerName="nova-manage" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429459 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-api" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429471 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" containerName="nova-api-log" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429482 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" containerName="nova-metadata-metadata" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.429488 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" containerName="nova-scheduler-scheduler" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.430388 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.436537 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.448072 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.463569 4770 scope.go:117] "RemoveContainer" containerID="f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e" Oct 04 04:38:17 crc kubenswrapper[4770]: E1004 04:38:17.464106 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e\": container with ID starting with f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e not found: ID does not exist" containerID="f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.464161 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e"} err="failed to get container status \"f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e\": rpc error: code = NotFound desc = could not find container \"f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e\": container with ID starting with f61511a9dcd9dfd925493792a2f8ba909451f5d92cede300373ce6f4f6749e9e not found: ID does not exist" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.464188 4770 scope.go:117] "RemoveContainer" containerID="ede9786b26fd9af96e9899c0d6294943617452f1e5ea820df101c54ca1a1e198" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.473046 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.486699 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.494323 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.495315 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bpld\" (UniqueName: \"kubernetes.io/projected/9fb764a3-bf4a-4838-8edd-ea89da1290ec-kube-api-access-2bpld\") pod \"nova-scheduler-0\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.495376 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-config-data\") pod \"nova-scheduler-0\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.495516 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.495993 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.496395 4770 scope.go:117] "RemoveContainer" containerID="72a4e783e08af8f179860f61d0026cebedbf48ebe9d9ec2186f5e449ab6635ab" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.499778 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.501323 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.504111 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.506584 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.511720 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.520595 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.596936 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bpld\" (UniqueName: \"kubernetes.io/projected/9fb764a3-bf4a-4838-8edd-ea89da1290ec-kube-api-access-2bpld\") pod \"nova-scheduler-0\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597025 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnvdx\" (UniqueName: \"kubernetes.io/projected/c11ee2b7-e83a-4100-8c9d-99614e71fab4-kube-api-access-xnvdx\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597060 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-config-data\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597087 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-config-data\") pod \"nova-scheduler-0\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597130 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11ee2b7-e83a-4100-8c9d-99614e71fab4-logs\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597162 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597196 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-config-data\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597232 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsnjd\" (UniqueName: \"kubernetes.io/projected/f8f24a90-871f-4527-a638-70353c990ac4-kube-api-access-nsnjd\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597278 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597311 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8f24a90-871f-4527-a638-70353c990ac4-logs\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.597973 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.601904 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-config-data\") pod \"nova-scheduler-0\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.602531 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.611835 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bpld\" (UniqueName: \"kubernetes.io/projected/9fb764a3-bf4a-4838-8edd-ea89da1290ec-kube-api-access-2bpld\") pod \"nova-scheduler-0\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.684599 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ab59ae1-24dc-49f6-b109-f12204ba4e38" path="/var/lib/kubelet/pods/2ab59ae1-24dc-49f6-b109-f12204ba4e38/volumes" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.685782 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16" path="/var/lib/kubelet/pods/58d5dd4d-f47a-4f2a-8a6b-cc81ffebde16/volumes" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.686499 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60006b51-98cb-4d97-8f4a-fb71c4e2b5fb" path="/var/lib/kubelet/pods/60006b51-98cb-4d97-8f4a-fb71c4e2b5fb/volumes" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.699700 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.699750 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8f24a90-871f-4527-a638-70353c990ac4-logs\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.699811 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.699844 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnvdx\" (UniqueName: \"kubernetes.io/projected/c11ee2b7-e83a-4100-8c9d-99614e71fab4-kube-api-access-xnvdx\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.699862 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-config-data\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.699923 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11ee2b7-e83a-4100-8c9d-99614e71fab4-logs\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.699955 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-config-data\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.699980 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsnjd\" (UniqueName: \"kubernetes.io/projected/f8f24a90-871f-4527-a638-70353c990ac4-kube-api-access-nsnjd\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.700255 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8f24a90-871f-4527-a638-70353c990ac4-logs\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.700831 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11ee2b7-e83a-4100-8c9d-99614e71fab4-logs\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.703368 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-config-data\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.704023 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-config-data\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.704883 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.705370 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.715858 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnvdx\" (UniqueName: \"kubernetes.io/projected/c11ee2b7-e83a-4100-8c9d-99614e71fab4-kube-api-access-xnvdx\") pod \"nova-metadata-0\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " pod="openstack/nova-metadata-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.717115 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsnjd\" (UniqueName: \"kubernetes.io/projected/f8f24a90-871f-4527-a638-70353c990ac4-kube-api-access-nsnjd\") pod \"nova-api-0\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.757523 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.817106 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:38:17 crc kubenswrapper[4770]: I1004 04:38:17.827391 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:38:18 crc kubenswrapper[4770]: I1004 04:38:18.251775 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:38:18 crc kubenswrapper[4770]: I1004 04:38:18.318782 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:38:18 crc kubenswrapper[4770]: W1004 04:38:18.319252 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8f24a90_871f_4527_a638_70353c990ac4.slice/crio-688d1ae84c51a56d9ffdf6a96baaf558c77bf7636eff6dd8e15a89f188c9e20c WatchSource:0}: Error finding container 688d1ae84c51a56d9ffdf6a96baaf558c77bf7636eff6dd8e15a89f188c9e20c: Status 404 returned error can't find the container with id 688d1ae84c51a56d9ffdf6a96baaf558c77bf7636eff6dd8e15a89f188c9e20c Oct 04 04:38:18 crc kubenswrapper[4770]: I1004 04:38:18.334962 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9fb764a3-bf4a-4838-8edd-ea89da1290ec","Type":"ContainerStarted","Data":"99b493523239646e019644fdc3cd6959643ed362d5f348e07bb6066c5d9d99f4"} Oct 04 04:38:18 crc kubenswrapper[4770]: I1004 04:38:18.338216 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f8f24a90-871f-4527-a638-70353c990ac4","Type":"ContainerStarted","Data":"688d1ae84c51a56d9ffdf6a96baaf558c77bf7636eff6dd8e15a89f188c9e20c"} Oct 04 04:38:18 crc kubenswrapper[4770]: W1004 04:38:18.373848 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc11ee2b7_e83a_4100_8c9d_99614e71fab4.slice/crio-6b3016fc7772c3f6c31817bad6dbc483ae366679dabd3f55174e24f46b37f136 WatchSource:0}: Error finding container 6b3016fc7772c3f6c31817bad6dbc483ae366679dabd3f55174e24f46b37f136: Status 404 returned error can't find the container with id 6b3016fc7772c3f6c31817bad6dbc483ae366679dabd3f55174e24f46b37f136 Oct 04 04:38:18 crc kubenswrapper[4770]: I1004 04:38:18.376721 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:38:19 crc kubenswrapper[4770]: I1004 04:38:19.353594 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f8f24a90-871f-4527-a638-70353c990ac4","Type":"ContainerStarted","Data":"4b1237dd4e92df8eaff2bc5c4dadefda8a7b322d72e597244a2ee6a39465a243"} Oct 04 04:38:19 crc kubenswrapper[4770]: I1004 04:38:19.353972 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f8f24a90-871f-4527-a638-70353c990ac4","Type":"ContainerStarted","Data":"78af8ac3916f0a131710ff1e44cb2da15bef5ebb6a701c99eff99f1ec819ce22"} Oct 04 04:38:19 crc kubenswrapper[4770]: I1004 04:38:19.355767 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9fb764a3-bf4a-4838-8edd-ea89da1290ec","Type":"ContainerStarted","Data":"5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3"} Oct 04 04:38:19 crc kubenswrapper[4770]: I1004 04:38:19.358555 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c11ee2b7-e83a-4100-8c9d-99614e71fab4","Type":"ContainerStarted","Data":"18610a114935514df2347c796285fa02ef9ad86635b0b5bba877f2449ce56634"} Oct 04 04:38:19 crc kubenswrapper[4770]: I1004 04:38:19.358610 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c11ee2b7-e83a-4100-8c9d-99614e71fab4","Type":"ContainerStarted","Data":"847616c491e6bfba61e03d3d93044508b49124d675d607457f62f7f96c14305f"} Oct 04 04:38:19 crc kubenswrapper[4770]: I1004 04:38:19.358623 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c11ee2b7-e83a-4100-8c9d-99614e71fab4","Type":"ContainerStarted","Data":"6b3016fc7772c3f6c31817bad6dbc483ae366679dabd3f55174e24f46b37f136"} Oct 04 04:38:19 crc kubenswrapper[4770]: I1004 04:38:19.377607 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.377586693 podStartE2EDuration="2.377586693s" podCreationTimestamp="2025-10-04 04:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:38:19.374929854 +0000 UTC m=+5710.666939586" watchObservedRunningTime="2025-10-04 04:38:19.377586693 +0000 UTC m=+5710.669596425" Oct 04 04:38:19 crc kubenswrapper[4770]: I1004 04:38:19.408837 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.40881649 podStartE2EDuration="2.40881649s" podCreationTimestamp="2025-10-04 04:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:38:19.393278723 +0000 UTC m=+5710.685288485" watchObservedRunningTime="2025-10-04 04:38:19.40881649 +0000 UTC m=+5710.700826192" Oct 04 04:38:19 crc kubenswrapper[4770]: I1004 04:38:19.419826 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.419805947 podStartE2EDuration="2.419805947s" podCreationTimestamp="2025-10-04 04:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:38:19.414057927 +0000 UTC m=+5710.706067689" watchObservedRunningTime="2025-10-04 04:38:19.419805947 +0000 UTC m=+5710.711815679" Oct 04 04:38:22 crc kubenswrapper[4770]: I1004 04:38:22.757724 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:38:22 crc kubenswrapper[4770]: I1004 04:38:22.828092 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:38:22 crc kubenswrapper[4770]: I1004 04:38:22.828179 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:38:27 crc kubenswrapper[4770]: I1004 04:38:27.758050 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:38:27 crc kubenswrapper[4770]: I1004 04:38:27.786156 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:38:27 crc kubenswrapper[4770]: I1004 04:38:27.817350 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:38:27 crc kubenswrapper[4770]: I1004 04:38:27.817410 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:38:27 crc kubenswrapper[4770]: I1004 04:38:27.828592 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:38:27 crc kubenswrapper[4770]: I1004 04:38:27.828637 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:38:28 crc kubenswrapper[4770]: I1004 04:38:28.503510 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:38:28 crc kubenswrapper[4770]: I1004 04:38:28.942197 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:38:28 crc kubenswrapper[4770]: I1004 04:38:28.942491 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:38:28 crc kubenswrapper[4770]: I1004 04:38:28.942526 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:38:28 crc kubenswrapper[4770]: I1004 04:38:28.942553 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:38:37 crc kubenswrapper[4770]: I1004 04:38:37.821301 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:38:37 crc kubenswrapper[4770]: I1004 04:38:37.821883 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:38:37 crc kubenswrapper[4770]: I1004 04:38:37.822539 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:38:37 crc kubenswrapper[4770]: I1004 04:38:37.822590 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:38:37 crc kubenswrapper[4770]: I1004 04:38:37.826264 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:38:37 crc kubenswrapper[4770]: I1004 04:38:37.827953 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:38:37 crc kubenswrapper[4770]: I1004 04:38:37.833217 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:38:37 crc kubenswrapper[4770]: I1004 04:38:37.835806 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:38:37 crc kubenswrapper[4770]: I1004 04:38:37.835956 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.043562 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bf68cc86f-8c9dw"] Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.045058 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.063427 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf68cc86f-8c9dw"] Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.176083 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.176237 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-config\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.176284 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.176364 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nqt6\" (UniqueName: \"kubernetes.io/projected/b90e62d6-09d2-4084-9762-2e2a5cef45dc-kube-api-access-5nqt6\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.176521 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-dns-svc\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.277828 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nqt6\" (UniqueName: \"kubernetes.io/projected/b90e62d6-09d2-4084-9762-2e2a5cef45dc-kube-api-access-5nqt6\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.277907 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-dns-svc\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.277953 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.278064 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-config\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.278092 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.279208 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.279272 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-dns-svc\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.279299 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-config\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.279326 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.302417 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nqt6\" (UniqueName: \"kubernetes.io/projected/b90e62d6-09d2-4084-9762-2e2a5cef45dc-kube-api-access-5nqt6\") pod \"dnsmasq-dns-6bf68cc86f-8c9dw\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.365637 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.547277 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:38:38 crc kubenswrapper[4770]: I1004 04:38:38.909904 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf68cc86f-8c9dw"] Oct 04 04:38:39 crc kubenswrapper[4770]: I1004 04:38:39.553269 4770 generic.go:334] "Generic (PLEG): container finished" podID="b90e62d6-09d2-4084-9762-2e2a5cef45dc" containerID="6a51f95acd2f96ec620189e82a3590543e173291f1fb0f675b7e4878664cb850" exitCode=0 Oct 04 04:38:39 crc kubenswrapper[4770]: I1004 04:38:39.553341 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" event={"ID":"b90e62d6-09d2-4084-9762-2e2a5cef45dc","Type":"ContainerDied","Data":"6a51f95acd2f96ec620189e82a3590543e173291f1fb0f675b7e4878664cb850"} Oct 04 04:38:39 crc kubenswrapper[4770]: I1004 04:38:39.553608 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" event={"ID":"b90e62d6-09d2-4084-9762-2e2a5cef45dc","Type":"ContainerStarted","Data":"6f203e85689eb9cfc81e69067b2d9194d20b30801f15e444edcf7a25f7606c62"} Oct 04 04:38:40 crc kubenswrapper[4770]: I1004 04:38:40.568826 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" event={"ID":"b90e62d6-09d2-4084-9762-2e2a5cef45dc","Type":"ContainerStarted","Data":"dfb8fc47ee82654121d34d69f42ec706252093c7c3ae63bbb1f7581212e8ab88"} Oct 04 04:38:40 crc kubenswrapper[4770]: I1004 04:38:40.591225 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" podStartSLOduration=2.591201443 podStartE2EDuration="2.591201443s" podCreationTimestamp="2025-10-04 04:38:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:38:40.588662377 +0000 UTC m=+5731.880672109" watchObservedRunningTime="2025-10-04 04:38:40.591201443 +0000 UTC m=+5731.883211185" Oct 04 04:38:41 crc kubenswrapper[4770]: I1004 04:38:41.583282 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:48 crc kubenswrapper[4770]: I1004 04:38:48.368327 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:38:48 crc kubenswrapper[4770]: I1004 04:38:48.436804 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7784f5b875-slvnh"] Oct 04 04:38:48 crc kubenswrapper[4770]: I1004 04:38:48.437122 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" podUID="facf4f26-1f71-4f73-9d66-aaeb7476628d" containerName="dnsmasq-dns" containerID="cri-o://820c7c61b51c718bcbe2424572a15765e125f7805069efba11a7265a35577e82" gracePeriod=10 Oct 04 04:38:48 crc kubenswrapper[4770]: I1004 04:38:48.668683 4770 generic.go:334] "Generic (PLEG): container finished" podID="facf4f26-1f71-4f73-9d66-aaeb7476628d" containerID="820c7c61b51c718bcbe2424572a15765e125f7805069efba11a7265a35577e82" exitCode=0 Oct 04 04:38:48 crc kubenswrapper[4770]: I1004 04:38:48.670532 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" event={"ID":"facf4f26-1f71-4f73-9d66-aaeb7476628d","Type":"ContainerDied","Data":"820c7c61b51c718bcbe2424572a15765e125f7805069efba11a7265a35577e82"} Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.520452 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.683281 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.697643 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7784f5b875-slvnh" event={"ID":"facf4f26-1f71-4f73-9d66-aaeb7476628d","Type":"ContainerDied","Data":"1fcbbda759a57d98340482a6357a339c96b1edee92ae7aca954a69a8de13753f"} Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.697743 4770 scope.go:117] "RemoveContainer" containerID="820c7c61b51c718bcbe2424572a15765e125f7805069efba11a7265a35577e82" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.712182 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-dns-svc\") pod \"facf4f26-1f71-4f73-9d66-aaeb7476628d\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.712256 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-sb\") pod \"facf4f26-1f71-4f73-9d66-aaeb7476628d\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.712341 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-nb\") pod \"facf4f26-1f71-4f73-9d66-aaeb7476628d\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.712392 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-config\") pod \"facf4f26-1f71-4f73-9d66-aaeb7476628d\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.712491 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk447\" (UniqueName: \"kubernetes.io/projected/facf4f26-1f71-4f73-9d66-aaeb7476628d-kube-api-access-nk447\") pod \"facf4f26-1f71-4f73-9d66-aaeb7476628d\" (UID: \"facf4f26-1f71-4f73-9d66-aaeb7476628d\") " Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.718910 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/facf4f26-1f71-4f73-9d66-aaeb7476628d-kube-api-access-nk447" (OuterVolumeSpecName: "kube-api-access-nk447") pod "facf4f26-1f71-4f73-9d66-aaeb7476628d" (UID: "facf4f26-1f71-4f73-9d66-aaeb7476628d"). InnerVolumeSpecName "kube-api-access-nk447". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.719250 4770 scope.go:117] "RemoveContainer" containerID="92171d8dbe7b6e0888d2f9ad95bbc962b7fb273c5a8868260336ae6e5d234b29" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.765316 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-config" (OuterVolumeSpecName: "config") pod "facf4f26-1f71-4f73-9d66-aaeb7476628d" (UID: "facf4f26-1f71-4f73-9d66-aaeb7476628d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.774194 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "facf4f26-1f71-4f73-9d66-aaeb7476628d" (UID: "facf4f26-1f71-4f73-9d66-aaeb7476628d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.783717 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "facf4f26-1f71-4f73-9d66-aaeb7476628d" (UID: "facf4f26-1f71-4f73-9d66-aaeb7476628d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.790163 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "facf4f26-1f71-4f73-9d66-aaeb7476628d" (UID: "facf4f26-1f71-4f73-9d66-aaeb7476628d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.814115 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.814291 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk447\" (UniqueName: \"kubernetes.io/projected/facf4f26-1f71-4f73-9d66-aaeb7476628d-kube-api-access-nk447\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.814372 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.814431 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:49 crc kubenswrapper[4770]: I1004 04:38:49.814523 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/facf4f26-1f71-4f73-9d66-aaeb7476628d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:50 crc kubenswrapper[4770]: I1004 04:38:50.016326 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7784f5b875-slvnh"] Oct 04 04:38:50 crc kubenswrapper[4770]: I1004 04:38:50.026567 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7784f5b875-slvnh"] Oct 04 04:38:51 crc kubenswrapper[4770]: I1004 04:38:51.689516 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="facf4f26-1f71-4f73-9d66-aaeb7476628d" path="/var/lib/kubelet/pods/facf4f26-1f71-4f73-9d66-aaeb7476628d/volumes" Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.211206 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-g7gcl"] Oct 04 04:38:52 crc kubenswrapper[4770]: E1004 04:38:52.211853 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="facf4f26-1f71-4f73-9d66-aaeb7476628d" containerName="init" Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.211871 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="facf4f26-1f71-4f73-9d66-aaeb7476628d" containerName="init" Oct 04 04:38:52 crc kubenswrapper[4770]: E1004 04:38:52.211895 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="facf4f26-1f71-4f73-9d66-aaeb7476628d" containerName="dnsmasq-dns" Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.211902 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="facf4f26-1f71-4f73-9d66-aaeb7476628d" containerName="dnsmasq-dns" Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.212253 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="facf4f26-1f71-4f73-9d66-aaeb7476628d" containerName="dnsmasq-dns" Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.213114 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g7gcl" Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.241846 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-g7gcl"] Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.363121 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29gx5\" (UniqueName: \"kubernetes.io/projected/b33454c9-e1a4-4239-a473-5b47223d090d-kube-api-access-29gx5\") pod \"cinder-db-create-g7gcl\" (UID: \"b33454c9-e1a4-4239-a473-5b47223d090d\") " pod="openstack/cinder-db-create-g7gcl" Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.464659 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29gx5\" (UniqueName: \"kubernetes.io/projected/b33454c9-e1a4-4239-a473-5b47223d090d-kube-api-access-29gx5\") pod \"cinder-db-create-g7gcl\" (UID: \"b33454c9-e1a4-4239-a473-5b47223d090d\") " pod="openstack/cinder-db-create-g7gcl" Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.484304 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29gx5\" (UniqueName: \"kubernetes.io/projected/b33454c9-e1a4-4239-a473-5b47223d090d-kube-api-access-29gx5\") pod \"cinder-db-create-g7gcl\" (UID: \"b33454c9-e1a4-4239-a473-5b47223d090d\") " pod="openstack/cinder-db-create-g7gcl" Oct 04 04:38:52 crc kubenswrapper[4770]: I1004 04:38:52.541875 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g7gcl" Oct 04 04:38:53 crc kubenswrapper[4770]: I1004 04:38:53.018699 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-g7gcl"] Oct 04 04:38:53 crc kubenswrapper[4770]: I1004 04:38:53.732940 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g7gcl" event={"ID":"b33454c9-e1a4-4239-a473-5b47223d090d","Type":"ContainerStarted","Data":"a97e14ae27fb18684c6f130fd5e5c7568cf392c4f90c5a078ffdad08e10fe1c1"} Oct 04 04:38:53 crc kubenswrapper[4770]: I1004 04:38:53.733318 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g7gcl" event={"ID":"b33454c9-e1a4-4239-a473-5b47223d090d","Type":"ContainerStarted","Data":"d87f3d691e5b4e6ef5b78d78d4b55c6cd73d342296b734a8829cdd857f23dafc"} Oct 04 04:38:53 crc kubenswrapper[4770]: I1004 04:38:53.748987 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-g7gcl" podStartSLOduration=1.748968316 podStartE2EDuration="1.748968316s" podCreationTimestamp="2025-10-04 04:38:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:38:53.748550104 +0000 UTC m=+5745.040559836" watchObservedRunningTime="2025-10-04 04:38:53.748968316 +0000 UTC m=+5745.040978058" Oct 04 04:38:54 crc kubenswrapper[4770]: I1004 04:38:54.746103 4770 generic.go:334] "Generic (PLEG): container finished" podID="b33454c9-e1a4-4239-a473-5b47223d090d" containerID="a97e14ae27fb18684c6f130fd5e5c7568cf392c4f90c5a078ffdad08e10fe1c1" exitCode=0 Oct 04 04:38:54 crc kubenswrapper[4770]: I1004 04:38:54.746256 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g7gcl" event={"ID":"b33454c9-e1a4-4239-a473-5b47223d090d","Type":"ContainerDied","Data":"a97e14ae27fb18684c6f130fd5e5c7568cf392c4f90c5a078ffdad08e10fe1c1"} Oct 04 04:38:56 crc kubenswrapper[4770]: I1004 04:38:56.096058 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g7gcl" Oct 04 04:38:56 crc kubenswrapper[4770]: I1004 04:38:56.228602 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29gx5\" (UniqueName: \"kubernetes.io/projected/b33454c9-e1a4-4239-a473-5b47223d090d-kube-api-access-29gx5\") pod \"b33454c9-e1a4-4239-a473-5b47223d090d\" (UID: \"b33454c9-e1a4-4239-a473-5b47223d090d\") " Oct 04 04:38:56 crc kubenswrapper[4770]: I1004 04:38:56.234667 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b33454c9-e1a4-4239-a473-5b47223d090d-kube-api-access-29gx5" (OuterVolumeSpecName: "kube-api-access-29gx5") pod "b33454c9-e1a4-4239-a473-5b47223d090d" (UID: "b33454c9-e1a4-4239-a473-5b47223d090d"). InnerVolumeSpecName "kube-api-access-29gx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:38:56 crc kubenswrapper[4770]: I1004 04:38:56.331254 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29gx5\" (UniqueName: \"kubernetes.io/projected/b33454c9-e1a4-4239-a473-5b47223d090d-kube-api-access-29gx5\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:56 crc kubenswrapper[4770]: I1004 04:38:56.765999 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-g7gcl" event={"ID":"b33454c9-e1a4-4239-a473-5b47223d090d","Type":"ContainerDied","Data":"d87f3d691e5b4e6ef5b78d78d4b55c6cd73d342296b734a8829cdd857f23dafc"} Oct 04 04:38:56 crc kubenswrapper[4770]: I1004 04:38:56.766066 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d87f3d691e5b4e6ef5b78d78d4b55c6cd73d342296b734a8829cdd857f23dafc" Oct 04 04:38:56 crc kubenswrapper[4770]: I1004 04:38:56.766155 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-g7gcl" Oct 04 04:39:01 crc kubenswrapper[4770]: I1004 04:39:01.796519 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:39:01 crc kubenswrapper[4770]: I1004 04:39:01.797108 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.272829 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-fdc9-account-create-j7fft"] Oct 04 04:39:02 crc kubenswrapper[4770]: E1004 04:39:02.273259 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33454c9-e1a4-4239-a473-5b47223d090d" containerName="mariadb-database-create" Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.273276 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33454c9-e1a4-4239-a473-5b47223d090d" containerName="mariadb-database-create" Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.273438 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b33454c9-e1a4-4239-a473-5b47223d090d" containerName="mariadb-database-create" Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.274103 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fdc9-account-create-j7fft" Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.276032 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.280295 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-fdc9-account-create-j7fft"] Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.439320 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r279\" (UniqueName: \"kubernetes.io/projected/b250070c-a70b-472c-9800-3a3797d57101-kube-api-access-4r279\") pod \"cinder-fdc9-account-create-j7fft\" (UID: \"b250070c-a70b-472c-9800-3a3797d57101\") " pod="openstack/cinder-fdc9-account-create-j7fft" Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.540743 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r279\" (UniqueName: \"kubernetes.io/projected/b250070c-a70b-472c-9800-3a3797d57101-kube-api-access-4r279\") pod \"cinder-fdc9-account-create-j7fft\" (UID: \"b250070c-a70b-472c-9800-3a3797d57101\") " pod="openstack/cinder-fdc9-account-create-j7fft" Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.559211 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r279\" (UniqueName: \"kubernetes.io/projected/b250070c-a70b-472c-9800-3a3797d57101-kube-api-access-4r279\") pod \"cinder-fdc9-account-create-j7fft\" (UID: \"b250070c-a70b-472c-9800-3a3797d57101\") " pod="openstack/cinder-fdc9-account-create-j7fft" Oct 04 04:39:02 crc kubenswrapper[4770]: I1004 04:39:02.622021 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fdc9-account-create-j7fft" Oct 04 04:39:03 crc kubenswrapper[4770]: I1004 04:39:03.116750 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-fdc9-account-create-j7fft"] Oct 04 04:39:03 crc kubenswrapper[4770]: W1004 04:39:03.124674 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb250070c_a70b_472c_9800_3a3797d57101.slice/crio-0b5120fff6992fd8d8e9fcda2836382043120609c8ffad00a5906da51f6454cb WatchSource:0}: Error finding container 0b5120fff6992fd8d8e9fcda2836382043120609c8ffad00a5906da51f6454cb: Status 404 returned error can't find the container with id 0b5120fff6992fd8d8e9fcda2836382043120609c8ffad00a5906da51f6454cb Oct 04 04:39:03 crc kubenswrapper[4770]: I1004 04:39:03.828485 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fdc9-account-create-j7fft" event={"ID":"b250070c-a70b-472c-9800-3a3797d57101","Type":"ContainerStarted","Data":"0b5120fff6992fd8d8e9fcda2836382043120609c8ffad00a5906da51f6454cb"} Oct 04 04:39:04 crc kubenswrapper[4770]: I1004 04:39:04.838817 4770 generic.go:334] "Generic (PLEG): container finished" podID="b250070c-a70b-472c-9800-3a3797d57101" containerID="29ffdc2a9210f2df9a291930a656a35612c13f93ab4ced3a39575d1362c0cece" exitCode=0 Oct 04 04:39:04 crc kubenswrapper[4770]: I1004 04:39:04.838873 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fdc9-account-create-j7fft" event={"ID":"b250070c-a70b-472c-9800-3a3797d57101","Type":"ContainerDied","Data":"29ffdc2a9210f2df9a291930a656a35612c13f93ab4ced3a39575d1362c0cece"} Oct 04 04:39:06 crc kubenswrapper[4770]: I1004 04:39:06.203870 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fdc9-account-create-j7fft" Oct 04 04:39:06 crc kubenswrapper[4770]: I1004 04:39:06.403156 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r279\" (UniqueName: \"kubernetes.io/projected/b250070c-a70b-472c-9800-3a3797d57101-kube-api-access-4r279\") pod \"b250070c-a70b-472c-9800-3a3797d57101\" (UID: \"b250070c-a70b-472c-9800-3a3797d57101\") " Oct 04 04:39:06 crc kubenswrapper[4770]: I1004 04:39:06.413848 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b250070c-a70b-472c-9800-3a3797d57101-kube-api-access-4r279" (OuterVolumeSpecName: "kube-api-access-4r279") pod "b250070c-a70b-472c-9800-3a3797d57101" (UID: "b250070c-a70b-472c-9800-3a3797d57101"). InnerVolumeSpecName "kube-api-access-4r279". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:06 crc kubenswrapper[4770]: I1004 04:39:06.506177 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r279\" (UniqueName: \"kubernetes.io/projected/b250070c-a70b-472c-9800-3a3797d57101-kube-api-access-4r279\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:06 crc kubenswrapper[4770]: I1004 04:39:06.856310 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fdc9-account-create-j7fft" event={"ID":"b250070c-a70b-472c-9800-3a3797d57101","Type":"ContainerDied","Data":"0b5120fff6992fd8d8e9fcda2836382043120609c8ffad00a5906da51f6454cb"} Oct 04 04:39:06 crc kubenswrapper[4770]: I1004 04:39:06.856354 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b5120fff6992fd8d8e9fcda2836382043120609c8ffad00a5906da51f6454cb" Oct 04 04:39:06 crc kubenswrapper[4770]: I1004 04:39:06.856747 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fdc9-account-create-j7fft" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.526908 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-pgbb9"] Oct 04 04:39:07 crc kubenswrapper[4770]: E1004 04:39:07.527455 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b250070c-a70b-472c-9800-3a3797d57101" containerName="mariadb-account-create" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.527474 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b250070c-a70b-472c-9800-3a3797d57101" containerName="mariadb-account-create" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.527709 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b250070c-a70b-472c-9800-3a3797d57101" containerName="mariadb-account-create" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.528484 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.530591 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.531253 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.531461 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bgrsf" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.535927 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pgbb9"] Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.730218 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-combined-ca-bundle\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.730443 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-scripts\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.730470 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-config-data\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.731071 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-etc-machine-id\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.731105 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-db-sync-config-data\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.731235 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvd48\" (UniqueName: \"kubernetes.io/projected/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-kube-api-access-lvd48\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.832809 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-scripts\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.833173 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-config-data\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.833262 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-etc-machine-id\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.833289 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-db-sync-config-data\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.833332 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvd48\" (UniqueName: \"kubernetes.io/projected/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-kube-api-access-lvd48\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.833385 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-etc-machine-id\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.833419 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-combined-ca-bundle\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.838729 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-scripts\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.838782 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-combined-ca-bundle\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.839037 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-config-data\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.843222 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-db-sync-config-data\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:07 crc kubenswrapper[4770]: I1004 04:39:07.850103 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvd48\" (UniqueName: \"kubernetes.io/projected/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-kube-api-access-lvd48\") pod \"cinder-db-sync-pgbb9\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:08 crc kubenswrapper[4770]: I1004 04:39:08.143962 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:08 crc kubenswrapper[4770]: I1004 04:39:08.594091 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pgbb9"] Oct 04 04:39:08 crc kubenswrapper[4770]: W1004 04:39:08.600817 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5f22307_4c2c_4c2b_82d6_6e323d4995e4.slice/crio-7891bb56ae7a370d3571b1db99256ba54bf36474258cc2eb02b68594f86e449c WatchSource:0}: Error finding container 7891bb56ae7a370d3571b1db99256ba54bf36474258cc2eb02b68594f86e449c: Status 404 returned error can't find the container with id 7891bb56ae7a370d3571b1db99256ba54bf36474258cc2eb02b68594f86e449c Oct 04 04:39:08 crc kubenswrapper[4770]: I1004 04:39:08.873579 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pgbb9" event={"ID":"e5f22307-4c2c-4c2b-82d6-6e323d4995e4","Type":"ContainerStarted","Data":"7891bb56ae7a370d3571b1db99256ba54bf36474258cc2eb02b68594f86e449c"} Oct 04 04:39:09 crc kubenswrapper[4770]: I1004 04:39:09.895656 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pgbb9" event={"ID":"e5f22307-4c2c-4c2b-82d6-6e323d4995e4","Type":"ContainerStarted","Data":"e196645e8452456154ccd1036a1638f33e40f698839d9a335c09babc302dc4a9"} Oct 04 04:39:09 crc kubenswrapper[4770]: I1004 04:39:09.919107 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-pgbb9" podStartSLOduration=2.91908937 podStartE2EDuration="2.91908937s" podCreationTimestamp="2025-10-04 04:39:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:39:09.912750624 +0000 UTC m=+5761.204760346" watchObservedRunningTime="2025-10-04 04:39:09.91908937 +0000 UTC m=+5761.211099082" Oct 04 04:39:31 crc kubenswrapper[4770]: I1004 04:39:31.795486 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:39:31 crc kubenswrapper[4770]: I1004 04:39:31.796326 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:39:43 crc kubenswrapper[4770]: I1004 04:39:43.258770 4770 generic.go:334] "Generic (PLEG): container finished" podID="e5f22307-4c2c-4c2b-82d6-6e323d4995e4" containerID="e196645e8452456154ccd1036a1638f33e40f698839d9a335c09babc302dc4a9" exitCode=0 Oct 04 04:39:43 crc kubenswrapper[4770]: I1004 04:39:43.258886 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pgbb9" event={"ID":"e5f22307-4c2c-4c2b-82d6-6e323d4995e4","Type":"ContainerDied","Data":"e196645e8452456154ccd1036a1638f33e40f698839d9a335c09babc302dc4a9"} Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.613272 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.768998 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-combined-ca-bundle\") pod \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.769280 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-config-data\") pod \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.769398 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-etc-machine-id\") pod \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.769454 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e5f22307-4c2c-4c2b-82d6-6e323d4995e4" (UID: "e5f22307-4c2c-4c2b-82d6-6e323d4995e4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.769601 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-db-sync-config-data\") pod \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.769734 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-scripts\") pod \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.769805 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvd48\" (UniqueName: \"kubernetes.io/projected/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-kube-api-access-lvd48\") pod \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\" (UID: \"e5f22307-4c2c-4c2b-82d6-6e323d4995e4\") " Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.770965 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.775860 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e5f22307-4c2c-4c2b-82d6-6e323d4995e4" (UID: "e5f22307-4c2c-4c2b-82d6-6e323d4995e4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.777057 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-kube-api-access-lvd48" (OuterVolumeSpecName: "kube-api-access-lvd48") pod "e5f22307-4c2c-4c2b-82d6-6e323d4995e4" (UID: "e5f22307-4c2c-4c2b-82d6-6e323d4995e4"). InnerVolumeSpecName "kube-api-access-lvd48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.777378 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-scripts" (OuterVolumeSpecName: "scripts") pod "e5f22307-4c2c-4c2b-82d6-6e323d4995e4" (UID: "e5f22307-4c2c-4c2b-82d6-6e323d4995e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.813465 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5f22307-4c2c-4c2b-82d6-6e323d4995e4" (UID: "e5f22307-4c2c-4c2b-82d6-6e323d4995e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.821361 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-config-data" (OuterVolumeSpecName: "config-data") pod "e5f22307-4c2c-4c2b-82d6-6e323d4995e4" (UID: "e5f22307-4c2c-4c2b-82d6-6e323d4995e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.872512 4770 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.872547 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.872556 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvd48\" (UniqueName: \"kubernetes.io/projected/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-kube-api-access-lvd48\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.872565 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:44 crc kubenswrapper[4770]: I1004 04:39:44.872576 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f22307-4c2c-4c2b-82d6-6e323d4995e4-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.280405 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pgbb9" event={"ID":"e5f22307-4c2c-4c2b-82d6-6e323d4995e4","Type":"ContainerDied","Data":"7891bb56ae7a370d3571b1db99256ba54bf36474258cc2eb02b68594f86e449c"} Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.280449 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7891bb56ae7a370d3571b1db99256ba54bf36474258cc2eb02b68594f86e449c" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.280494 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pgbb9" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.649067 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c66885869-nwzqr"] Oct 04 04:39:45 crc kubenswrapper[4770]: E1004 04:39:45.649773 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f22307-4c2c-4c2b-82d6-6e323d4995e4" containerName="cinder-db-sync" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.649786 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f22307-4c2c-4c2b-82d6-6e323d4995e4" containerName="cinder-db-sync" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.649970 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f22307-4c2c-4c2b-82d6-6e323d4995e4" containerName="cinder-db-sync" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.650945 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.756621 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c66885869-nwzqr"] Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.787365 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.788876 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.792583 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.792821 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.792966 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bgrsf" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.793153 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.797189 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.832835 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-config\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.832898 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-scripts\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.832958 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.832995 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch9vd\" (UniqueName: \"kubernetes.io/projected/6a94bbc1-1520-4811-b035-e812395b31a1-kube-api-access-ch9vd\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.833171 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klncv\" (UniqueName: \"kubernetes.io/projected/99081418-459d-4d41-9d13-447f409f1d8f-kube-api-access-klncv\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.833228 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-sb\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.833414 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.833646 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99081418-459d-4d41-9d13-447f409f1d8f-logs\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.833685 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-nb\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.833830 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99081418-459d-4d41-9d13-447f409f1d8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.833867 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-dns-svc\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.833896 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.935680 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99081418-459d-4d41-9d13-447f409f1d8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.935798 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99081418-459d-4d41-9d13-447f409f1d8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.935842 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-dns-svc\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.935873 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936039 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-config\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936114 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-scripts\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936202 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936309 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch9vd\" (UniqueName: \"kubernetes.io/projected/6a94bbc1-1520-4811-b035-e812395b31a1-kube-api-access-ch9vd\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936349 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klncv\" (UniqueName: \"kubernetes.io/projected/99081418-459d-4d41-9d13-447f409f1d8f-kube-api-access-klncv\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936374 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-sb\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936460 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936600 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99081418-459d-4d41-9d13-447f409f1d8f-logs\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936625 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-nb\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.936879 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-dns-svc\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.937418 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-nb\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.938913 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99081418-459d-4d41-9d13-447f409f1d8f-logs\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.939163 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-sb\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.939515 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-config\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.941535 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.942674 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.947338 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.952552 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-scripts\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.954545 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klncv\" (UniqueName: \"kubernetes.io/projected/99081418-459d-4d41-9d13-447f409f1d8f-kube-api-access-klncv\") pod \"cinder-api-0\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " pod="openstack/cinder-api-0" Oct 04 04:39:45 crc kubenswrapper[4770]: I1004 04:39:45.959366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch9vd\" (UniqueName: \"kubernetes.io/projected/6a94bbc1-1520-4811-b035-e812395b31a1-kube-api-access-ch9vd\") pod \"dnsmasq-dns-5c66885869-nwzqr\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:46 crc kubenswrapper[4770]: I1004 04:39:46.020732 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:46 crc kubenswrapper[4770]: I1004 04:39:46.118862 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:39:46 crc kubenswrapper[4770]: I1004 04:39:46.512053 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c66885869-nwzqr"] Oct 04 04:39:46 crc kubenswrapper[4770]: I1004 04:39:46.688927 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:39:46 crc kubenswrapper[4770]: W1004 04:39:46.704131 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99081418_459d_4d41_9d13_447f409f1d8f.slice/crio-449f5de8def159c36f2529874350c4fa2ee87a74052313f0b79fa2e8bb9e65ac WatchSource:0}: Error finding container 449f5de8def159c36f2529874350c4fa2ee87a74052313f0b79fa2e8bb9e65ac: Status 404 returned error can't find the container with id 449f5de8def159c36f2529874350c4fa2ee87a74052313f0b79fa2e8bb9e65ac Oct 04 04:39:47 crc kubenswrapper[4770]: I1004 04:39:47.306277 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99081418-459d-4d41-9d13-447f409f1d8f","Type":"ContainerStarted","Data":"449f5de8def159c36f2529874350c4fa2ee87a74052313f0b79fa2e8bb9e65ac"} Oct 04 04:39:47 crc kubenswrapper[4770]: I1004 04:39:47.307777 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" event={"ID":"6a94bbc1-1520-4811-b035-e812395b31a1","Type":"ContainerStarted","Data":"2537f03ad69d69d119ae0d33630a3e794fe8a004ad722e801026ef7c22da7135"} Oct 04 04:39:47 crc kubenswrapper[4770]: I1004 04:39:47.307796 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" event={"ID":"6a94bbc1-1520-4811-b035-e812395b31a1","Type":"ContainerStarted","Data":"c5f03a128ab433fd913e8d9aac8747b2ccf6c209e9ca61b53cfaeaee6995a0b0"} Oct 04 04:39:48 crc kubenswrapper[4770]: I1004 04:39:48.320516 4770 generic.go:334] "Generic (PLEG): container finished" podID="6a94bbc1-1520-4811-b035-e812395b31a1" containerID="2537f03ad69d69d119ae0d33630a3e794fe8a004ad722e801026ef7c22da7135" exitCode=0 Oct 04 04:39:48 crc kubenswrapper[4770]: I1004 04:39:48.320639 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" event={"ID":"6a94bbc1-1520-4811-b035-e812395b31a1","Type":"ContainerDied","Data":"2537f03ad69d69d119ae0d33630a3e794fe8a004ad722e801026ef7c22da7135"} Oct 04 04:39:48 crc kubenswrapper[4770]: I1004 04:39:48.324359 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99081418-459d-4d41-9d13-447f409f1d8f","Type":"ContainerStarted","Data":"bfe37593f19384cca39e89138fb81f3d89bccb96202c8b8910482b4d51e07959"} Oct 04 04:39:49 crc kubenswrapper[4770]: I1004 04:39:49.335294 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" event={"ID":"6a94bbc1-1520-4811-b035-e812395b31a1","Type":"ContainerStarted","Data":"2bed80eb41587a7333fdced63965c7b12a05d662454d30d64046d65f9a23f6eb"} Oct 04 04:39:49 crc kubenswrapper[4770]: I1004 04:39:49.335680 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:49 crc kubenswrapper[4770]: I1004 04:39:49.338188 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99081418-459d-4d41-9d13-447f409f1d8f","Type":"ContainerStarted","Data":"ee9ef5ebc010bc349dc626a5f4a465ae9408be0d70fdb133d406dba8ec69bb1a"} Oct 04 04:39:49 crc kubenswrapper[4770]: I1004 04:39:49.338442 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 04:39:49 crc kubenswrapper[4770]: I1004 04:39:49.369485 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" podStartSLOduration=4.369462136 podStartE2EDuration="4.369462136s" podCreationTimestamp="2025-10-04 04:39:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:39:49.359043884 +0000 UTC m=+5800.651053596" watchObservedRunningTime="2025-10-04 04:39:49.369462136 +0000 UTC m=+5800.661471848" Oct 04 04:39:49 crc kubenswrapper[4770]: I1004 04:39:49.382758 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.382733033 podStartE2EDuration="4.382733033s" podCreationTimestamp="2025-10-04 04:39:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:39:49.377173917 +0000 UTC m=+5800.669183629" watchObservedRunningTime="2025-10-04 04:39:49.382733033 +0000 UTC m=+5800.674742765" Oct 04 04:39:56 crc kubenswrapper[4770]: I1004 04:39:56.023246 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:39:56 crc kubenswrapper[4770]: I1004 04:39:56.099428 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf68cc86f-8c9dw"] Oct 04 04:39:56 crc kubenswrapper[4770]: I1004 04:39:56.100288 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" podUID="b90e62d6-09d2-4084-9762-2e2a5cef45dc" containerName="dnsmasq-dns" containerID="cri-o://dfb8fc47ee82654121d34d69f42ec706252093c7c3ae63bbb1f7581212e8ab88" gracePeriod=10 Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.419655 4770 generic.go:334] "Generic (PLEG): container finished" podID="b90e62d6-09d2-4084-9762-2e2a5cef45dc" containerID="dfb8fc47ee82654121d34d69f42ec706252093c7c3ae63bbb1f7581212e8ab88" exitCode=0 Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.419736 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" event={"ID":"b90e62d6-09d2-4084-9762-2e2a5cef45dc","Type":"ContainerDied","Data":"dfb8fc47ee82654121d34d69f42ec706252093c7c3ae63bbb1f7581212e8ab88"} Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.420084 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" event={"ID":"b90e62d6-09d2-4084-9762-2e2a5cef45dc","Type":"ContainerDied","Data":"6f203e85689eb9cfc81e69067b2d9194d20b30801f15e444edcf7a25f7606c62"} Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.420105 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f203e85689eb9cfc81e69067b2d9194d20b30801f15e444edcf7a25f7606c62" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.443584 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.616194 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-nb\") pod \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.616252 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-sb\") pod \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.616282 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-config\") pod \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.616359 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-dns-svc\") pod \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.616525 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nqt6\" (UniqueName: \"kubernetes.io/projected/b90e62d6-09d2-4084-9762-2e2a5cef45dc-kube-api-access-5nqt6\") pod \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\" (UID: \"b90e62d6-09d2-4084-9762-2e2a5cef45dc\") " Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.624556 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b90e62d6-09d2-4084-9762-2e2a5cef45dc-kube-api-access-5nqt6" (OuterVolumeSpecName: "kube-api-access-5nqt6") pod "b90e62d6-09d2-4084-9762-2e2a5cef45dc" (UID: "b90e62d6-09d2-4084-9762-2e2a5cef45dc"). InnerVolumeSpecName "kube-api-access-5nqt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.662250 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b90e62d6-09d2-4084-9762-2e2a5cef45dc" (UID: "b90e62d6-09d2-4084-9762-2e2a5cef45dc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.703285 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-config" (OuterVolumeSpecName: "config") pod "b90e62d6-09d2-4084-9762-2e2a5cef45dc" (UID: "b90e62d6-09d2-4084-9762-2e2a5cef45dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.708699 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b90e62d6-09d2-4084-9762-2e2a5cef45dc" (UID: "b90e62d6-09d2-4084-9762-2e2a5cef45dc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.712562 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b90e62d6-09d2-4084-9762-2e2a5cef45dc" (UID: "b90e62d6-09d2-4084-9762-2e2a5cef45dc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.719967 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.720047 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.720066 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.720085 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b90e62d6-09d2-4084-9762-2e2a5cef45dc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:57 crc kubenswrapper[4770]: I1004 04:39:57.720102 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nqt6\" (UniqueName: \"kubernetes.io/projected/b90e62d6-09d2-4084-9762-2e2a5cef45dc-kube-api-access-5nqt6\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:58 crc kubenswrapper[4770]: I1004 04:39:58.119244 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 04:39:58 crc kubenswrapper[4770]: I1004 04:39:58.428703 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf68cc86f-8c9dw" Oct 04 04:39:58 crc kubenswrapper[4770]: I1004 04:39:58.458988 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf68cc86f-8c9dw"] Oct 04 04:39:58 crc kubenswrapper[4770]: I1004 04:39:58.466739 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bf68cc86f-8c9dw"] Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.417525 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.417979 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" containerID="cri-o://d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" gracePeriod=30 Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.440078 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.440324 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-log" containerID="cri-o://78af8ac3916f0a131710ff1e44cb2da15bef5ebb6a701c99eff99f1ec819ce22" gracePeriod=30 Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.440714 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-api" containerID="cri-o://4b1237dd4e92df8eaff2bc5c4dadefda8a7b322d72e597244a2ee6a39465a243" gracePeriod=30 Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.456197 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.456389 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" containerID="cri-o://5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" gracePeriod=30 Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.468470 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.468640 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://1513f37950192789203ff20cde5e658cc3e08a222bcc4798c8e0bf256043e3e5" gracePeriod=30 Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.478604 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.478844 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" containerID="cri-o://847616c491e6bfba61e03d3d93044508b49124d675d607457f62f7f96c14305f" gracePeriod=30 Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.479265 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" containerID="cri-o://18610a114935514df2347c796285fa02ef9ad86635b0b5bba877f2449ce56634" gracePeriod=30 Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.595235 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.595452 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" containerID="cri-o://20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" gracePeriod=30 Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.683734 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b90e62d6-09d2-4084-9762-2e2a5cef45dc" path="/var/lib/kubelet/pods/b90e62d6-09d2-4084-9762-2e2a5cef45dc/volumes" Oct 04 04:39:59 crc kubenswrapper[4770]: I1004 04:39:59.731408 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:00 crc kubenswrapper[4770]: I1004 04:40:00.450447 4770 generic.go:334] "Generic (PLEG): container finished" podID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerID="847616c491e6bfba61e03d3d93044508b49124d675d607457f62f7f96c14305f" exitCode=143 Oct 04 04:40:00 crc kubenswrapper[4770]: I1004 04:40:00.450513 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c11ee2b7-e83a-4100-8c9d-99614e71fab4","Type":"ContainerDied","Data":"847616c491e6bfba61e03d3d93044508b49124d675d607457f62f7f96c14305f"} Oct 04 04:40:00 crc kubenswrapper[4770]: I1004 04:40:00.453523 4770 generic.go:334] "Generic (PLEG): container finished" podID="f8f24a90-871f-4527-a638-70353c990ac4" containerID="78af8ac3916f0a131710ff1e44cb2da15bef5ebb6a701c99eff99f1ec819ce22" exitCode=143 Oct 04 04:40:00 crc kubenswrapper[4770]: I1004 04:40:00.453589 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f8f24a90-871f-4527-a638-70353c990ac4","Type":"ContainerDied","Data":"78af8ac3916f0a131710ff1e44cb2da15bef5ebb6a701c99eff99f1ec819ce22"} Oct 04 04:40:00 crc kubenswrapper[4770]: I1004 04:40:00.455176 4770 generic.go:334] "Generic (PLEG): container finished" podID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerID="1513f37950192789203ff20cde5e658cc3e08a222bcc4798c8e0bf256043e3e5" exitCode=0 Oct 04 04:40:00 crc kubenswrapper[4770]: I1004 04:40:00.455214 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"39e02f97-9bb3-43b4-a2f8-7633252519a3","Type":"ContainerDied","Data":"1513f37950192789203ff20cde5e658cc3e08a222bcc4798c8e0bf256043e3e5"} Oct 04 04:40:01 crc kubenswrapper[4770]: I1004 04:40:01.796124 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:40:01 crc kubenswrapper[4770]: I1004 04:40:01.797389 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:40:01 crc kubenswrapper[4770]: I1004 04:40:01.797648 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:40:01 crc kubenswrapper[4770]: I1004 04:40:01.798863 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e30f9d379ccb3500c52f46d443a65b8cc328fea98419caf04a4e9a30d4bd0cb"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:40:01 crc kubenswrapper[4770]: I1004 04:40:01.799246 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://2e30f9d379ccb3500c52f46d443a65b8cc328fea98419caf04a4e9a30d4bd0cb" gracePeriod=600 Oct 04 04:40:02 crc kubenswrapper[4770]: E1004 04:40:02.352488 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:02 crc kubenswrapper[4770]: E1004 04:40:02.359276 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:02 crc kubenswrapper[4770]: E1004 04:40:02.360612 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:02 crc kubenswrapper[4770]: E1004 04:40:02.360678 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:02 crc kubenswrapper[4770]: E1004 04:40:02.760086 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:02 crc kubenswrapper[4770]: E1004 04:40:02.762376 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:02 crc kubenswrapper[4770]: E1004 04:40:02.763744 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:02 crc kubenswrapper[4770]: E1004 04:40:02.763790 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:02 crc kubenswrapper[4770]: I1004 04:40:02.828841 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:02 crc kubenswrapper[4770]: I1004 04:40:02.828886 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:03 crc kubenswrapper[4770]: E1004 04:40:03.405604 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:03 crc kubenswrapper[4770]: E1004 04:40:03.407307 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:03 crc kubenswrapper[4770]: E1004 04:40:03.409707 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:03 crc kubenswrapper[4770]: E1004 04:40:03.409827 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:04 crc kubenswrapper[4770]: I1004 04:40:04.058222 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-qhmd4"] Oct 04 04:40:04 crc kubenswrapper[4770]: I1004 04:40:04.069331 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-qhmd4"] Oct 04 04:40:04 crc kubenswrapper[4770]: I1004 04:40:04.731929 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:05 crc kubenswrapper[4770]: I1004 04:40:05.696492 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147d8093-1179-48bb-a771-8cb1b6a7f8a7" path="/var/lib/kubelet/pods/147d8093-1179-48bb-a771-8cb1b6a7f8a7/volumes" Oct 04 04:40:07 crc kubenswrapper[4770]: E1004 04:40:07.354704 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:07 crc kubenswrapper[4770]: E1004 04:40:07.358383 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:07 crc kubenswrapper[4770]: E1004 04:40:07.365593 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:07 crc kubenswrapper[4770]: E1004 04:40:07.365715 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:07 crc kubenswrapper[4770]: E1004 04:40:07.760609 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:07 crc kubenswrapper[4770]: E1004 04:40:07.762767 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:07 crc kubenswrapper[4770]: E1004 04:40:07.764590 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:07 crc kubenswrapper[4770]: E1004 04:40:07.764662 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:07 crc kubenswrapper[4770]: I1004 04:40:07.828817 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:07 crc kubenswrapper[4770]: I1004 04:40:07.828825 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:08 crc kubenswrapper[4770]: E1004 04:40:08.406168 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:08 crc kubenswrapper[4770]: E1004 04:40:08.412720 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:08 crc kubenswrapper[4770]: E1004 04:40:08.414935 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:08 crc kubenswrapper[4770]: E1004 04:40:08.415051 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:09 crc kubenswrapper[4770]: I1004 04:40:09.731384 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:09 crc kubenswrapper[4770]: I1004 04:40:09.731509 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:40:10 crc kubenswrapper[4770]: I1004 04:40:10.265734 4770 generic.go:334] "Generic (PLEG): container finished" podID="f8f24a90-871f-4527-a638-70353c990ac4" containerID="4b1237dd4e92df8eaff2bc5c4dadefda8a7b322d72e597244a2ee6a39465a243" exitCode=-1 Oct 04 04:40:10 crc kubenswrapper[4770]: I1004 04:40:10.266082 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f8f24a90-871f-4527-a638-70353c990ac4","Type":"ContainerDied","Data":"4b1237dd4e92df8eaff2bc5c4dadefda8a7b322d72e597244a2ee6a39465a243"} Oct 04 04:40:12 crc kubenswrapper[4770]: E1004 04:40:12.353329 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:12 crc kubenswrapper[4770]: E1004 04:40:12.356587 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:12 crc kubenswrapper[4770]: E1004 04:40:12.359083 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:12 crc kubenswrapper[4770]: E1004 04:40:12.359233 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:12 crc kubenswrapper[4770]: E1004 04:40:12.760596 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:12 crc kubenswrapper[4770]: E1004 04:40:12.763817 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:12 crc kubenswrapper[4770]: E1004 04:40:12.766144 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:12 crc kubenswrapper[4770]: E1004 04:40:12.766420 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:12 crc kubenswrapper[4770]: I1004 04:40:12.827831 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:12 crc kubenswrapper[4770]: I1004 04:40:12.827971 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:40:12 crc kubenswrapper[4770]: I1004 04:40:12.827978 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:12 crc kubenswrapper[4770]: I1004 04:40:12.828111 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:40:13 crc kubenswrapper[4770]: E1004 04:40:13.405207 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:13 crc kubenswrapper[4770]: E1004 04:40:13.406845 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:13 crc kubenswrapper[4770]: E1004 04:40:13.408632 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:13 crc kubenswrapper[4770]: E1004 04:40:13.408693 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:14 crc kubenswrapper[4770]: I1004 04:40:14.037403 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-44b0-account-create-cpdz8"] Oct 04 04:40:14 crc kubenswrapper[4770]: I1004 04:40:14.056239 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-44b0-account-create-cpdz8"] Oct 04 04:40:14 crc kubenswrapper[4770]: I1004 04:40:14.731476 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:15 crc kubenswrapper[4770]: I1004 04:40:15.690353 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7101c35-8518-43e9-8a1f-fed84491a6dc" path="/var/lib/kubelet/pods/a7101c35-8518-43e9-8a1f-fed84491a6dc/volumes" Oct 04 04:40:17 crc kubenswrapper[4770]: E1004 04:40:17.353100 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:17 crc kubenswrapper[4770]: E1004 04:40:17.355851 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:17 crc kubenswrapper[4770]: E1004 04:40:17.358996 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:17 crc kubenswrapper[4770]: E1004 04:40:17.359088 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:17 crc kubenswrapper[4770]: E1004 04:40:17.761105 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:17 crc kubenswrapper[4770]: E1004 04:40:17.763492 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:17 crc kubenswrapper[4770]: E1004 04:40:17.766194 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:17 crc kubenswrapper[4770]: E1004 04:40:17.766261 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:17 crc kubenswrapper[4770]: I1004 04:40:17.819789 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": dial tcp 10.217.1.73:8774: connect: connection refused" Oct 04 04:40:17 crc kubenswrapper[4770]: I1004 04:40:17.821599 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": dial tcp 10.217.1.73:8774: connect: connection refused" Oct 04 04:40:17 crc kubenswrapper[4770]: I1004 04:40:17.829134 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:17 crc kubenswrapper[4770]: I1004 04:40:17.829200 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:18 crc kubenswrapper[4770]: E1004 04:40:18.405615 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:18 crc kubenswrapper[4770]: E1004 04:40:18.407635 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:18 crc kubenswrapper[4770]: E1004 04:40:18.410510 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:18 crc kubenswrapper[4770]: E1004 04:40:18.410625 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:19 crc kubenswrapper[4770]: I1004 04:40:19.731349 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:21 crc kubenswrapper[4770]: I1004 04:40:21.045717 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-dbmp8"] Oct 04 04:40:21 crc kubenswrapper[4770]: I1004 04:40:21.059267 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-dbmp8"] Oct 04 04:40:21 crc kubenswrapper[4770]: I1004 04:40:21.710172 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3274a0e-8276-49f0-9583-ef9bfdb43c6e" path="/var/lib/kubelet/pods/c3274a0e-8276-49f0-9583-ef9bfdb43c6e/volumes" Oct 04 04:40:22 crc kubenswrapper[4770]: E1004 04:40:22.353489 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:22 crc kubenswrapper[4770]: E1004 04:40:22.356236 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:22 crc kubenswrapper[4770]: E1004 04:40:22.358829 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:22 crc kubenswrapper[4770]: E1004 04:40:22.358876 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:22 crc kubenswrapper[4770]: E1004 04:40:22.761592 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:22 crc kubenswrapper[4770]: E1004 04:40:22.770815 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:22 crc kubenswrapper[4770]: E1004 04:40:22.776665 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:22 crc kubenswrapper[4770]: E1004 04:40:22.776752 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:22 crc kubenswrapper[4770]: I1004 04:40:22.828117 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:22 crc kubenswrapper[4770]: I1004 04:40:22.828677 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:23 crc kubenswrapper[4770]: E1004 04:40:23.405706 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:23 crc kubenswrapper[4770]: E1004 04:40:23.407705 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:23 crc kubenswrapper[4770]: E1004 04:40:23.409918 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:23 crc kubenswrapper[4770]: E1004 04:40:23.409959 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:24 crc kubenswrapper[4770]: I1004 04:40:24.731148 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:25 crc kubenswrapper[4770]: I1004 04:40:25.554442 4770 scope.go:117] "RemoveContainer" containerID="06f2ed82f10c96ad2ef5cf70956895ed9d0d0c77a294c20c2ed0adb7c6f8dfa9" Oct 04 04:40:25 crc kubenswrapper[4770]: I1004 04:40:25.596418 4770 scope.go:117] "RemoveContainer" containerID="bb45933b28e6a908156865440b73265b4f3d6c8febdcece1962dcda742904783" Oct 04 04:40:25 crc kubenswrapper[4770]: I1004 04:40:25.636872 4770 scope.go:117] "RemoveContainer" containerID="1312cd478db70cdc0313cd3744c6a773b954d9be4c176178827f9ca822adf4bc" Oct 04 04:40:27 crc kubenswrapper[4770]: E1004 04:40:27.354140 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:27 crc kubenswrapper[4770]: E1004 04:40:27.356857 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:27 crc kubenswrapper[4770]: E1004 04:40:27.359047 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:27 crc kubenswrapper[4770]: E1004 04:40:27.359125 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:27 crc kubenswrapper[4770]: E1004 04:40:27.760983 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:27 crc kubenswrapper[4770]: E1004 04:40:27.763179 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:27 crc kubenswrapper[4770]: E1004 04:40:27.765358 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:27 crc kubenswrapper[4770]: E1004 04:40:27.765439 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:27 crc kubenswrapper[4770]: I1004 04:40:27.828645 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:27 crc kubenswrapper[4770]: I1004 04:40:27.828699 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:28 crc kubenswrapper[4770]: E1004 04:40:28.406503 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:28 crc kubenswrapper[4770]: E1004 04:40:28.409061 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:28 crc kubenswrapper[4770]: E1004 04:40:28.411884 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:28 crc kubenswrapper[4770]: E1004 04:40:28.411990 4770 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:29 crc kubenswrapper[4770]: I1004 04:40:29.730607 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:32 crc kubenswrapper[4770]: E1004 04:40:32.351659 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:32 crc kubenswrapper[4770]: E1004 04:40:32.352760 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:32 crc kubenswrapper[4770]: E1004 04:40:32.353156 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:32 crc kubenswrapper[4770]: E1004 04:40:32.353243 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:32 crc kubenswrapper[4770]: E1004 04:40:32.759294 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:32 crc kubenswrapper[4770]: E1004 04:40:32.759825 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:32 crc kubenswrapper[4770]: E1004 04:40:32.760591 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:32 crc kubenswrapper[4770]: E1004 04:40:32.760704 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:32 crc kubenswrapper[4770]: I1004 04:40:32.828223 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:32 crc kubenswrapper[4770]: I1004 04:40:32.835591 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:33 crc kubenswrapper[4770]: E1004 04:40:33.403871 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:33 crc kubenswrapper[4770]: E1004 04:40:33.404317 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:33 crc kubenswrapper[4770]: E1004 04:40:33.404781 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:33 crc kubenswrapper[4770]: E1004 04:40:33.404864 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:34 crc kubenswrapper[4770]: I1004 04:40:34.732484 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:35 crc kubenswrapper[4770]: I1004 04:40:35.050834 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fkn6b"] Oct 04 04:40:35 crc kubenswrapper[4770]: I1004 04:40:35.066241 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fkn6b"] Oct 04 04:40:35 crc kubenswrapper[4770]: I1004 04:40:35.694703 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3706ecc-dda1-498f-8319-b10379ceccb9" path="/var/lib/kubelet/pods/a3706ecc-dda1-498f-8319-b10379ceccb9/volumes" Oct 04 04:40:37 crc kubenswrapper[4770]: E1004 04:40:37.351422 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:37 crc kubenswrapper[4770]: E1004 04:40:37.352426 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:37 crc kubenswrapper[4770]: E1004 04:40:37.353355 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:37 crc kubenswrapper[4770]: E1004 04:40:37.353449 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:37 crc kubenswrapper[4770]: E1004 04:40:37.758614 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:37 crc kubenswrapper[4770]: E1004 04:40:37.759447 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:37 crc kubenswrapper[4770]: E1004 04:40:37.759947 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:37 crc kubenswrapper[4770]: E1004 04:40:37.760061 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:37 crc kubenswrapper[4770]: I1004 04:40:37.829336 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:37 crc kubenswrapper[4770]: I1004 04:40:37.829492 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:38 crc kubenswrapper[4770]: E1004 04:40:38.403521 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:38 crc kubenswrapper[4770]: E1004 04:40:38.404264 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:38 crc kubenswrapper[4770]: E1004 04:40:38.404891 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:38 crc kubenswrapper[4770]: E1004 04:40:38.404945 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:39 crc kubenswrapper[4770]: I1004 04:40:39.731026 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:42 crc kubenswrapper[4770]: E1004 04:40:42.351958 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:42 crc kubenswrapper[4770]: E1004 04:40:42.353484 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:42 crc kubenswrapper[4770]: E1004 04:40:42.353998 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:42 crc kubenswrapper[4770]: E1004 04:40:42.354124 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:42 crc kubenswrapper[4770]: E1004 04:40:42.758357 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:42 crc kubenswrapper[4770]: E1004 04:40:42.759137 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:42 crc kubenswrapper[4770]: E1004 04:40:42.759760 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:42 crc kubenswrapper[4770]: E1004 04:40:42.759861 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:42 crc kubenswrapper[4770]: I1004 04:40:42.827848 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:42 crc kubenswrapper[4770]: I1004 04:40:42.828240 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:43 crc kubenswrapper[4770]: E1004 04:40:43.403677 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:43 crc kubenswrapper[4770]: E1004 04:40:43.404356 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:43 crc kubenswrapper[4770]: E1004 04:40:43.404920 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:43 crc kubenswrapper[4770]: E1004 04:40:43.404995 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:44 crc kubenswrapper[4770]: I1004 04:40:44.732674 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:47 crc kubenswrapper[4770]: E1004 04:40:47.351465 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:47 crc kubenswrapper[4770]: E1004 04:40:47.352149 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:47 crc kubenswrapper[4770]: E1004 04:40:47.352403 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:47 crc kubenswrapper[4770]: E1004 04:40:47.352424 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:47 crc kubenswrapper[4770]: E1004 04:40:47.759181 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:47 crc kubenswrapper[4770]: E1004 04:40:47.760385 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:47 crc kubenswrapper[4770]: E1004 04:40:47.761074 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:47 crc kubenswrapper[4770]: E1004 04:40:47.761144 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:47 crc kubenswrapper[4770]: I1004 04:40:47.819633 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": dial tcp 10.217.1.73:8774: connect: connection refused" Oct 04 04:40:47 crc kubenswrapper[4770]: I1004 04:40:47.819646 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": dial tcp 10.217.1.73:8774: connect: connection refused" Oct 04 04:40:47 crc kubenswrapper[4770]: I1004 04:40:47.829424 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:47 crc kubenswrapper[4770]: I1004 04:40:47.829486 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:48 crc kubenswrapper[4770]: E1004 04:40:48.403876 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:48 crc kubenswrapper[4770]: E1004 04:40:48.404660 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:48 crc kubenswrapper[4770]: E1004 04:40:48.405557 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:48 crc kubenswrapper[4770]: E1004 04:40:48.405621 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:49 crc kubenswrapper[4770]: I1004 04:40:49.731496 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:52 crc kubenswrapper[4770]: E1004 04:40:52.351192 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:52 crc kubenswrapper[4770]: E1004 04:40:52.352488 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:52 crc kubenswrapper[4770]: E1004 04:40:52.352925 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:52 crc kubenswrapper[4770]: E1004 04:40:52.353041 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:52 crc kubenswrapper[4770]: E1004 04:40:52.759557 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:52 crc kubenswrapper[4770]: E1004 04:40:52.760044 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:52 crc kubenswrapper[4770]: E1004 04:40:52.760551 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:52 crc kubenswrapper[4770]: E1004 04:40:52.760639 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:52 crc kubenswrapper[4770]: I1004 04:40:52.829675 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:52 crc kubenswrapper[4770]: I1004 04:40:52.829684 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:53 crc kubenswrapper[4770]: E1004 04:40:53.407776 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:53 crc kubenswrapper[4770]: E1004 04:40:53.408762 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:53 crc kubenswrapper[4770]: E1004 04:40:53.409537 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:53 crc kubenswrapper[4770]: E1004 04:40:53.409656 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:54 crc kubenswrapper[4770]: I1004 04:40:54.731285 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:40:57 crc kubenswrapper[4770]: E1004 04:40:57.350781 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:57 crc kubenswrapper[4770]: E1004 04:40:57.351806 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:57 crc kubenswrapper[4770]: E1004 04:40:57.355440 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:57 crc kubenswrapper[4770]: E1004 04:40:57.355531 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:40:57 crc kubenswrapper[4770]: E1004 04:40:57.759580 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:57 crc kubenswrapper[4770]: E1004 04:40:57.760214 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:57 crc kubenswrapper[4770]: E1004 04:40:57.760739 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:40:57 crc kubenswrapper[4770]: E1004 04:40:57.760859 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:40:57 crc kubenswrapper[4770]: I1004 04:40:57.828823 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:57 crc kubenswrapper[4770]: I1004 04:40:57.828887 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": dial tcp 10.217.1.74:8775: connect: connection refused" Oct 04 04:40:58 crc kubenswrapper[4770]: E1004 04:40:58.404032 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:58 crc kubenswrapper[4770]: E1004 04:40:58.404865 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:58 crc kubenswrapper[4770]: E1004 04:40:58.405565 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:40:58 crc kubenswrapper[4770]: E1004 04:40:58.405670 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:40:59 crc kubenswrapper[4770]: I1004 04:40:59.731421 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.64:6080/vnc_lite.html\": dial tcp 10.217.1.64:6080: connect: connection refused" Oct 04 04:41:01 crc kubenswrapper[4770]: I1004 04:41:01.600388 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_machine-config-daemon-h6msx_6bd3be93-1791-4cd1-b3ae-b4032548e93a/machine-config-daemon/18.log" Oct 04 04:41:01 crc kubenswrapper[4770]: I1004 04:41:01.603875 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="2e30f9d379ccb3500c52f46d443a65b8cc328fea98419caf04a4e9a30d4bd0cb" exitCode=-1 Oct 04 04:41:01 crc kubenswrapper[4770]: I1004 04:41:01.603984 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"2e30f9d379ccb3500c52f46d443a65b8cc328fea98419caf04a4e9a30d4bd0cb"} Oct 04 04:41:01 crc kubenswrapper[4770]: I1004 04:41:01.604118 4770 scope.go:117] "RemoveContainer" containerID="4fe64471279ab07a1c58f08c2c5cef3c813f8d149dac7fbb66fd5a319c47cbe5" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.138514 4770 generic.go:334] "Generic (PLEG): container finished" podID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerID="18610a114935514df2347c796285fa02ef9ad86635b0b5bba877f2449ce56634" exitCode=0 Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.138616 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c11ee2b7-e83a-4100-8c9d-99614e71fab4","Type":"ContainerDied","Data":"18610a114935514df2347c796285fa02ef9ad86635b0b5bba877f2449ce56634"} Oct 04 04:41:02 crc kubenswrapper[4770]: E1004 04:41:02.358651 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:41:02 crc kubenswrapper[4770]: E1004 04:41:02.359024 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:41:02 crc kubenswrapper[4770]: E1004 04:41:02.359282 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 04:41:02 crc kubenswrapper[4770]: E1004 04:41:02.359318 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.383638 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.505773 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-combined-ca-bundle\") pod \"39e02f97-9bb3-43b4-a2f8-7633252519a3\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.505977 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-config-data\") pod \"39e02f97-9bb3-43b4-a2f8-7633252519a3\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.506002 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/39e02f97-9bb3-43b4-a2f8-7633252519a3-kube-api-access-hbjp4\") pod \"39e02f97-9bb3-43b4-a2f8-7633252519a3\" (UID: \"39e02f97-9bb3-43b4-a2f8-7633252519a3\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.520949 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39e02f97-9bb3-43b4-a2f8-7633252519a3-kube-api-access-hbjp4" (OuterVolumeSpecName: "kube-api-access-hbjp4") pod "39e02f97-9bb3-43b4-a2f8-7633252519a3" (UID: "39e02f97-9bb3-43b4-a2f8-7633252519a3"). InnerVolumeSpecName "kube-api-access-hbjp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.547114 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39e02f97-9bb3-43b4-a2f8-7633252519a3" (UID: "39e02f97-9bb3-43b4-a2f8-7633252519a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.558980 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-config-data" (OuterVolumeSpecName: "config-data") pod "39e02f97-9bb3-43b4-a2f8-7633252519a3" (UID: "39e02f97-9bb3-43b4-a2f8-7633252519a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.608955 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.616885 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.621506 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.621533 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/39e02f97-9bb3-43b4-a2f8-7633252519a3-kube-api-access-hbjp4\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.621551 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e02f97-9bb3-43b4-a2f8-7633252519a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.627067 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.648980 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.651435 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.724590 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnvdx\" (UniqueName: \"kubernetes.io/projected/c11ee2b7-e83a-4100-8c9d-99614e71fab4-kube-api-access-xnvdx\") pod \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.724668 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11ee2b7-e83a-4100-8c9d-99614e71fab4-logs\") pod \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.724721 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bpld\" (UniqueName: \"kubernetes.io/projected/9fb764a3-bf4a-4838-8edd-ea89da1290ec-kube-api-access-2bpld\") pod \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.724920 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv5hn\" (UniqueName: \"kubernetes.io/projected/50830e0d-eeca-4ee8-ad7a-adca578cb88f-kube-api-access-qv5hn\") pod \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.724975 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-combined-ca-bundle\") pod \"048a0b66-f6e3-424c-80c5-552afd14ed64\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725021 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsnjd\" (UniqueName: \"kubernetes.io/projected/f8f24a90-871f-4527-a638-70353c990ac4-kube-api-access-nsnjd\") pod \"f8f24a90-871f-4527-a638-70353c990ac4\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725063 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-config-data\") pod \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725097 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-combined-ca-bundle\") pod \"f8f24a90-871f-4527-a638-70353c990ac4\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725122 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-config-data\") pod \"f8f24a90-871f-4527-a638-70353c990ac4\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725220 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8f24a90-871f-4527-a638-70353c990ac4-logs\") pod \"f8f24a90-871f-4527-a638-70353c990ac4\" (UID: \"f8f24a90-871f-4527-a638-70353c990ac4\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725261 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdndj\" (UniqueName: \"kubernetes.io/projected/048a0b66-f6e3-424c-80c5-552afd14ed64-kube-api-access-hdndj\") pod \"048a0b66-f6e3-424c-80c5-552afd14ed64\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725309 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-config-data\") pod \"048a0b66-f6e3-424c-80c5-552afd14ed64\" (UID: \"048a0b66-f6e3-424c-80c5-552afd14ed64\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725351 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-config-data\") pod \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725382 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-config-data\") pod \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725408 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-combined-ca-bundle\") pod \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\" (UID: \"9fb764a3-bf4a-4838-8edd-ea89da1290ec\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725448 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-combined-ca-bundle\") pod \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\" (UID: \"50830e0d-eeca-4ee8-ad7a-adca578cb88f\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.725476 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-combined-ca-bundle\") pod \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\" (UID: \"c11ee2b7-e83a-4100-8c9d-99614e71fab4\") " Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.727778 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8f24a90-871f-4527-a638-70353c990ac4-logs" (OuterVolumeSpecName: "logs") pod "f8f24a90-871f-4527-a638-70353c990ac4" (UID: "f8f24a90-871f-4527-a638-70353c990ac4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.729795 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c11ee2b7-e83a-4100-8c9d-99614e71fab4-kube-api-access-xnvdx" (OuterVolumeSpecName: "kube-api-access-xnvdx") pod "c11ee2b7-e83a-4100-8c9d-99614e71fab4" (UID: "c11ee2b7-e83a-4100-8c9d-99614e71fab4"). InnerVolumeSpecName "kube-api-access-xnvdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.736028 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb764a3-bf4a-4838-8edd-ea89da1290ec-kube-api-access-2bpld" (OuterVolumeSpecName: "kube-api-access-2bpld") pod "9fb764a3-bf4a-4838-8edd-ea89da1290ec" (UID: "9fb764a3-bf4a-4838-8edd-ea89da1290ec"). InnerVolumeSpecName "kube-api-access-2bpld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.736466 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c11ee2b7-e83a-4100-8c9d-99614e71fab4-logs" (OuterVolumeSpecName: "logs") pod "c11ee2b7-e83a-4100-8c9d-99614e71fab4" (UID: "c11ee2b7-e83a-4100-8c9d-99614e71fab4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.763342 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048a0b66-f6e3-424c-80c5-552afd14ed64-kube-api-access-hdndj" (OuterVolumeSpecName: "kube-api-access-hdndj") pod "048a0b66-f6e3-424c-80c5-552afd14ed64" (UID: "048a0b66-f6e3-424c-80c5-552afd14ed64"). InnerVolumeSpecName "kube-api-access-hdndj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.764374 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f24a90-871f-4527-a638-70353c990ac4-kube-api-access-nsnjd" (OuterVolumeSpecName: "kube-api-access-nsnjd") pod "f8f24a90-871f-4527-a638-70353c990ac4" (UID: "f8f24a90-871f-4527-a638-70353c990ac4"). InnerVolumeSpecName "kube-api-access-nsnjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.772356 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50830e0d-eeca-4ee8-ad7a-adca578cb88f-kube-api-access-qv5hn" (OuterVolumeSpecName: "kube-api-access-qv5hn") pod "50830e0d-eeca-4ee8-ad7a-adca578cb88f" (UID: "50830e0d-eeca-4ee8-ad7a-adca578cb88f"). InnerVolumeSpecName "kube-api-access-qv5hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.814786 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c11ee2b7-e83a-4100-8c9d-99614e71fab4" (UID: "c11ee2b7-e83a-4100-8c9d-99614e71fab4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.818738 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fb764a3-bf4a-4838-8edd-ea89da1290ec" (UID: "9fb764a3-bf4a-4838-8edd-ea89da1290ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.819275 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-config-data" (OuterVolumeSpecName: "config-data") pod "9fb764a3-bf4a-4838-8edd-ea89da1290ec" (UID: "9fb764a3-bf4a-4838-8edd-ea89da1290ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832217 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832255 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832268 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnvdx\" (UniqueName: \"kubernetes.io/projected/c11ee2b7-e83a-4100-8c9d-99614e71fab4-kube-api-access-xnvdx\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832281 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11ee2b7-e83a-4100-8c9d-99614e71fab4-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832292 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bpld\" (UniqueName: \"kubernetes.io/projected/9fb764a3-bf4a-4838-8edd-ea89da1290ec-kube-api-access-2bpld\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832302 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv5hn\" (UniqueName: \"kubernetes.io/projected/50830e0d-eeca-4ee8-ad7a-adca578cb88f-kube-api-access-qv5hn\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832312 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsnjd\" (UniqueName: \"kubernetes.io/projected/f8f24a90-871f-4527-a638-70353c990ac4-kube-api-access-nsnjd\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832321 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8f24a90-871f-4527-a638-70353c990ac4-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832331 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdndj\" (UniqueName: \"kubernetes.io/projected/048a0b66-f6e3-424c-80c5-552afd14ed64-kube-api-access-hdndj\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.832342 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb764a3-bf4a-4838-8edd-ea89da1290ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.834197 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-config-data" (OuterVolumeSpecName: "config-data") pod "048a0b66-f6e3-424c-80c5-552afd14ed64" (UID: "048a0b66-f6e3-424c-80c5-552afd14ed64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.839266 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50830e0d-eeca-4ee8-ad7a-adca578cb88f" (UID: "50830e0d-eeca-4ee8-ad7a-adca578cb88f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.843125 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-config-data" (OuterVolumeSpecName: "config-data") pod "50830e0d-eeca-4ee8-ad7a-adca578cb88f" (UID: "50830e0d-eeca-4ee8-ad7a-adca578cb88f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.851372 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "048a0b66-f6e3-424c-80c5-552afd14ed64" (UID: "048a0b66-f6e3-424c-80c5-552afd14ed64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.856319 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-config-data" (OuterVolumeSpecName: "config-data") pod "c11ee2b7-e83a-4100-8c9d-99614e71fab4" (UID: "c11ee2b7-e83a-4100-8c9d-99614e71fab4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.880221 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-config-data" (OuterVolumeSpecName: "config-data") pod "f8f24a90-871f-4527-a638-70353c990ac4" (UID: "f8f24a90-871f-4527-a638-70353c990ac4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.890360 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8f24a90-871f-4527-a638-70353c990ac4" (UID: "f8f24a90-871f-4527-a638-70353c990ac4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.936480 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.936572 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f24a90-871f-4527-a638-70353c990ac4-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.936586 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.936596 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11ee2b7-e83a-4100-8c9d-99614e71fab4-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.936657 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.936670 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048a0b66-f6e3-424c-80c5-552afd14ed64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:02 crc kubenswrapper[4770]: I1004 04:41:02.936680 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50830e0d-eeca-4ee8-ad7a-adca578cb88f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.166202 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.176259 4770 generic.go:334] "Generic (PLEG): container finished" podID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" exitCode=137 Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.176346 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"048a0b66-f6e3-424c-80c5-552afd14ed64","Type":"ContainerDied","Data":"d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.176380 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"048a0b66-f6e3-424c-80c5-552afd14ed64","Type":"ContainerDied","Data":"3c5ce60c052f33b56d736d2c60294911bed95f56d15c06d8d62b5a1bbe496be8"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.176404 4770 scope.go:117] "RemoveContainer" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.176532 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.204171 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.204194 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c11ee2b7-e83a-4100-8c9d-99614e71fab4","Type":"ContainerDied","Data":"6b3016fc7772c3f6c31817bad6dbc483ae366679dabd3f55174e24f46b37f136"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.206330 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"39e02f97-9bb3-43b4-a2f8-7633252519a3","Type":"ContainerDied","Data":"2292442289fe31fec881436081a7ccc3450c73e61cc4c800213cfa4e04ee8caa"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.206431 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.213549 4770 generic.go:334] "Generic (PLEG): container finished" podID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" exitCode=137 Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.213620 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"50830e0d-eeca-4ee8-ad7a-adca578cb88f","Type":"ContainerDied","Data":"20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.213653 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"50830e0d-eeca-4ee8-ad7a-adca578cb88f","Type":"ContainerDied","Data":"c4f8165225bcb96b751a096646c2c6025942311f56f5385d05d3653376e3b325"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.213710 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.235448 4770 generic.go:334] "Generic (PLEG): container finished" podID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" exitCode=137 Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.235634 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.235628 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9fb764a3-bf4a-4838-8edd-ea89da1290ec","Type":"ContainerDied","Data":"5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.235795 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9fb764a3-bf4a-4838-8edd-ea89da1290ec","Type":"ContainerDied","Data":"99b493523239646e019644fdc3cd6959643ed362d5f348e07bb6066c5d9d99f4"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.243847 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f8f24a90-871f-4527-a638-70353c990ac4","Type":"ContainerDied","Data":"688d1ae84c51a56d9ffdf6a96baaf558c77bf7636eff6dd8e15a89f188c9e20c"} Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.243985 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.246215 4770 scope.go:117] "RemoveContainer" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.247186 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b\": container with ID starting with d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b not found: ID does not exist" containerID="d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.247224 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b"} err="failed to get container status \"d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b\": rpc error: code = NotFound desc = could not find container \"d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b\": container with ID starting with d7107d9dc22be89f4745477f84e729de73475bf0b341329c3b28ebde5e0e970b not found: ID does not exist" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.247247 4770 scope.go:117] "RemoveContainer" containerID="18610a114935514df2347c796285fa02ef9ad86635b0b5bba877f2449ce56634" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.283130 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.293239 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.313524 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.313962 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.313986 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.313999 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-api" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314008 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-api" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.314050 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90e62d6-09d2-4084-9762-2e2a5cef45dc" containerName="dnsmasq-dns" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314060 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90e62d6-09d2-4084-9762-2e2a5cef45dc" containerName="dnsmasq-dns" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.314074 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314082 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.314095 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90e62d6-09d2-4084-9762-2e2a5cef45dc" containerName="init" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314102 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90e62d6-09d2-4084-9762-2e2a5cef45dc" containerName="init" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.314118 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314126 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.314145 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314151 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.314162 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314169 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.314180 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314186 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.314195 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-log" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314200 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-log" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314373 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" containerName="nova-cell1-conductor-conductor" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314398 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" containerName="nova-cell0-conductor-conductor" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314416 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90e62d6-09d2-4084-9762-2e2a5cef45dc" containerName="dnsmasq-dns" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314429 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-api" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314441 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314456 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" containerName="nova-scheduler-scheduler" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314468 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-metadata" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314480 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f24a90-871f-4527-a638-70353c990ac4" containerName="nova-api-log" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.314494 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" containerName="nova-metadata-log" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.315212 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.318728 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.318886 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fclmn" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.324068 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.348307 4770 scope.go:117] "RemoveContainer" containerID="847616c491e6bfba61e03d3d93044508b49124d675d607457f62f7f96c14305f" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.410718 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.440053 4770 scope.go:117] "RemoveContainer" containerID="1513f37950192789203ff20cde5e658cc3e08a222bcc4798c8e0bf256043e3e5" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.450195 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzscp\" (UniqueName: \"kubernetes.io/projected/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-kube-api-access-tzscp\") pod \"nova-cell0-conductor-0\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.450262 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.450328 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.452900 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.468172 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.499741 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.502275 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.503458 4770 scope.go:117] "RemoveContainer" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.511214 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.511462 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.532104 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.537100 4770 scope.go:117] "RemoveContainer" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.537702 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b\": container with ID starting with 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b not found: ID does not exist" containerID="20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.537738 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b"} err="failed to get container status \"20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b\": rpc error: code = NotFound desc = could not find container \"20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b\": container with ID starting with 20f90471de2a6b93d8d07ddccdcd43e8f4f81aa5c549e4cede9aca207d663d7b not found: ID does not exist" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.537765 4770 scope.go:117] "RemoveContainer" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.541894 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.543377 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.558504 4770 scope.go:117] "RemoveContainer" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" Oct 04 04:41:03 crc kubenswrapper[4770]: E1004 04:41:03.559520 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3\": container with ID starting with 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 not found: ID does not exist" containerID="5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.559587 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3"} err="failed to get container status \"5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3\": rpc error: code = NotFound desc = could not find container \"5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3\": container with ID starting with 5b91b437d2875ba19c92d15caba147fce2fbfa5a36258dcf4b831bf72edb97f3 not found: ID does not exist" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.559612 4770 scope.go:117] "RemoveContainer" containerID="4b1237dd4e92df8eaff2bc5c4dadefda8a7b322d72e597244a2ee6a39465a243" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.570684 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.573089 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzscp\" (UniqueName: \"kubernetes.io/projected/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-kube-api-access-tzscp\") pod \"nova-cell0-conductor-0\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.573182 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.573326 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.584575 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.594809 4770 scope.go:117] "RemoveContainer" containerID="78af8ac3916f0a131710ff1e44cb2da15bef5ebb6a701c99eff99f1ec819ce22" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.603849 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.605006 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.607394 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzscp\" (UniqueName: \"kubernetes.io/projected/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-kube-api-access-tzscp\") pod \"nova-cell0-conductor-0\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.614067 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.629200 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.640662 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.651408 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.665445 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.666018 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.668049 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.671456 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.675181 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9czd\" (UniqueName: \"kubernetes.io/projected/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-kube-api-access-t9czd\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.675271 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.675301 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rx49\" (UniqueName: \"kubernetes.io/projected/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-kube-api-access-5rx49\") pod \"nova-cell1-conductor-0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.675335 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-config-data\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.675371 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-logs\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.675448 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.675472 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.692067 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="048a0b66-f6e3-424c-80c5-552afd14ed64" path="/var/lib/kubelet/pods/048a0b66-f6e3-424c-80c5-552afd14ed64/volumes" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.692769 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50830e0d-eeca-4ee8-ad7a-adca578cb88f" path="/var/lib/kubelet/pods/50830e0d-eeca-4ee8-ad7a-adca578cb88f/volumes" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.693404 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fb764a3-bf4a-4838-8edd-ea89da1290ec" path="/var/lib/kubelet/pods/9fb764a3-bf4a-4838-8edd-ea89da1290ec/volumes" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.696090 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c11ee2b7-e83a-4100-8c9d-99614e71fab4" path="/var/lib/kubelet/pods/c11ee2b7-e83a-4100-8c9d-99614e71fab4/volumes" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.696996 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8f24a90-871f-4527-a638-70353c990ac4" path="/var/lib/kubelet/pods/f8f24a90-871f-4527-a638-70353c990ac4/volumes" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.697799 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.697842 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.706017 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.708106 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.710843 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.719109 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.733043 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.739681 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.740846 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.743221 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.763210 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.778800 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9czd\" (UniqueName: \"kubernetes.io/projected/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-kube-api-access-t9czd\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.778898 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.778950 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.778982 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rx49\" (UniqueName: \"kubernetes.io/projected/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-kube-api-access-5rx49\") pod \"nova-cell1-conductor-0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.779053 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-config-data\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.779080 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-config-data\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.779133 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-logs\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.779210 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59068af-fca9-4d62-ba15-7f7306a32ae0-logs\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.779276 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxzvl\" (UniqueName: \"kubernetes.io/projected/b59068af-fca9-4d62-ba15-7f7306a32ae0-kube-api-access-fxzvl\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.779296 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.779314 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.784139 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.786123 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-logs\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.789238 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.790676 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-config-data\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.808901 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.811566 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rx49\" (UniqueName: \"kubernetes.io/projected/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-kube-api-access-5rx49\") pod \"nova-cell1-conductor-0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.816910 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9czd\" (UniqueName: \"kubernetes.io/projected/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-kube-api-access-t9czd\") pod \"nova-api-0\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.823047 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.878993 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.880678 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.880750 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/481d5341-bd64-43ad-be7d-54ce70e96f29-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"481d5341-bd64-43ad-be7d-54ce70e96f29\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.881703 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gtms\" (UniqueName: \"kubernetes.io/projected/481d5341-bd64-43ad-be7d-54ce70e96f29-kube-api-access-8gtms\") pod \"nova-cell1-novncproxy-0\" (UID: \"481d5341-bd64-43ad-be7d-54ce70e96f29\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.881874 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-config-data\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.882052 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59068af-fca9-4d62-ba15-7f7306a32ae0-logs\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.882145 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64h79\" (UniqueName: \"kubernetes.io/projected/46e7a275-1a83-4ee4-bbee-819afd9e1df6-kube-api-access-64h79\") pod \"nova-scheduler-0\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " pod="openstack/nova-scheduler-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.882201 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/481d5341-bd64-43ad-be7d-54ce70e96f29-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"481d5341-bd64-43ad-be7d-54ce70e96f29\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.882298 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxzvl\" (UniqueName: \"kubernetes.io/projected/b59068af-fca9-4d62-ba15-7f7306a32ae0-kube-api-access-fxzvl\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.882465 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-config-data\") pod \"nova-scheduler-0\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " pod="openstack/nova-scheduler-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.882500 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " pod="openstack/nova-scheduler-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.883039 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59068af-fca9-4d62-ba15-7f7306a32ae0-logs\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.885727 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-config-data\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.894200 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.906275 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxzvl\" (UniqueName: \"kubernetes.io/projected/b59068af-fca9-4d62-ba15-7f7306a32ae0-kube-api-access-fxzvl\") pod \"nova-metadata-0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " pod="openstack/nova-metadata-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.984426 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-config-data\") pod \"nova-scheduler-0\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " pod="openstack/nova-scheduler-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.984467 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " pod="openstack/nova-scheduler-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.984512 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/481d5341-bd64-43ad-be7d-54ce70e96f29-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"481d5341-bd64-43ad-be7d-54ce70e96f29\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.984546 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gtms\" (UniqueName: \"kubernetes.io/projected/481d5341-bd64-43ad-be7d-54ce70e96f29-kube-api-access-8gtms\") pod \"nova-cell1-novncproxy-0\" (UID: \"481d5341-bd64-43ad-be7d-54ce70e96f29\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.984611 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64h79\" (UniqueName: \"kubernetes.io/projected/46e7a275-1a83-4ee4-bbee-819afd9e1df6-kube-api-access-64h79\") pod \"nova-scheduler-0\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " pod="openstack/nova-scheduler-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.984629 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/481d5341-bd64-43ad-be7d-54ce70e96f29-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"481d5341-bd64-43ad-be7d-54ce70e96f29\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:03 crc kubenswrapper[4770]: I1004 04:41:03.988619 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/481d5341-bd64-43ad-be7d-54ce70e96f29-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"481d5341-bd64-43ad-be7d-54ce70e96f29\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.002278 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/481d5341-bd64-43ad-be7d-54ce70e96f29-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"481d5341-bd64-43ad-be7d-54ce70e96f29\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.002371 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-config-data\") pod \"nova-scheduler-0\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " pod="openstack/nova-scheduler-0" Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.004357 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " pod="openstack/nova-scheduler-0" Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.005317 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64h79\" (UniqueName: \"kubernetes.io/projected/46e7a275-1a83-4ee4-bbee-819afd9e1df6-kube-api-access-64h79\") pod \"nova-scheduler-0\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " pod="openstack/nova-scheduler-0" Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.008268 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gtms\" (UniqueName: \"kubernetes.io/projected/481d5341-bd64-43ad-be7d-54ce70e96f29-kube-api-access-8gtms\") pod \"nova-cell1-novncproxy-0\" (UID: \"481d5341-bd64-43ad-be7d-54ce70e96f29\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.084917 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.107799 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.112926 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.125557 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.273770 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.311555 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"6eceb699-4221-46eb-9b62-8ae2cf60cd5b","Type":"ContainerStarted","Data":"d0c0ad0ee77b6bd2213ee3ef438a3b31557fc30be183f588241e17f90c732e40"} Oct 04 04:41:04 crc kubenswrapper[4770]: W1004 04:41:04.321720 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78da1490_9b4b_442a_bb3f_0ad05e1dbf04.slice/crio-1367a090f87a30ea7101871c1d7d58d4a3ca79f66dbf43f1908fae718d3bae28 WatchSource:0}: Error finding container 1367a090f87a30ea7101871c1d7d58d4a3ca79f66dbf43f1908fae718d3bae28: Status 404 returned error can't find the container with id 1367a090f87a30ea7101871c1d7d58d4a3ca79f66dbf43f1908fae718d3bae28 Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.349486 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.625162 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:41:04 crc kubenswrapper[4770]: W1004 04:41:04.628704 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb59068af_fca9_4d62_ba15_7f7306a32ae0.slice/crio-7b695814a33d8874da6ca761c52f4b75ac4e102287f87d7f15e3455a90af64eb WatchSource:0}: Error finding container 7b695814a33d8874da6ca761c52f4b75ac4e102287f87d7f15e3455a90af64eb: Status 404 returned error can't find the container with id 7b695814a33d8874da6ca761c52f4b75ac4e102287f87d7f15e3455a90af64eb Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.673571 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:41:04 crc kubenswrapper[4770]: I1004 04:41:04.691398 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:41:04 crc kubenswrapper[4770]: W1004 04:41:04.710887 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46e7a275_1a83_4ee4_bbee_819afd9e1df6.slice/crio-55d139908fc565872075917de4d0a292f7ea2528f96553b97e4dba033641718b WatchSource:0}: Error finding container 55d139908fc565872075917de4d0a292f7ea2528f96553b97e4dba033641718b: Status 404 returned error can't find the container with id 55d139908fc565872075917de4d0a292f7ea2528f96553b97e4dba033641718b Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.354523 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b59068af-fca9-4d62-ba15-7f7306a32ae0","Type":"ContainerStarted","Data":"f60d1eeb54bc17f640f15c972d2bb137d7e0a8e712981795e91597a9ddc441b8"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.354799 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b59068af-fca9-4d62-ba15-7f7306a32ae0","Type":"ContainerStarted","Data":"099d516172bf35064ad73e6e674f0abde980b6c7dba602194a147839a424eddd"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.354826 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b59068af-fca9-4d62-ba15-7f7306a32ae0","Type":"ContainerStarted","Data":"7b695814a33d8874da6ca761c52f4b75ac4e102287f87d7f15e3455a90af64eb"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.356320 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"6eceb699-4221-46eb-9b62-8ae2cf60cd5b","Type":"ContainerStarted","Data":"bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.356497 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.357521 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0","Type":"ContainerStarted","Data":"608b9cfe609d2917d7ab9daf7942e98dba655ca0c09309610a6944f7b5d12fb7"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.357543 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0","Type":"ContainerStarted","Data":"47645a3176d23e317c2a5febfcd448a03051e6e523c6d3ef8e31cef9045111f8"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.357667 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.359239 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78da1490-9b4b-442a-bb3f-0ad05e1dbf04","Type":"ContainerStarted","Data":"67ea243dcc9ab412f13757ffb58194953795759d6551db8bd466f9ad8a7d83ca"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.359280 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78da1490-9b4b-442a-bb3f-0ad05e1dbf04","Type":"ContainerStarted","Data":"aac27019eb2828d08ce64585131c37f09c142e7a091370cca0513e33c3524167"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.359296 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78da1490-9b4b-442a-bb3f-0ad05e1dbf04","Type":"ContainerStarted","Data":"1367a090f87a30ea7101871c1d7d58d4a3ca79f66dbf43f1908fae718d3bae28"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.360641 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"481d5341-bd64-43ad-be7d-54ce70e96f29","Type":"ContainerStarted","Data":"75c131fa6160a34fa24dd46552b3bd42fa191e49480cc971fb8725481795dbc7"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.360683 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"481d5341-bd64-43ad-be7d-54ce70e96f29","Type":"ContainerStarted","Data":"c645eda0feccfa3a2c3cedfbeb411cbb2b6d6e4c10946b8822466e612a63129d"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.361950 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"46e7a275-1a83-4ee4-bbee-819afd9e1df6","Type":"ContainerStarted","Data":"a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.361983 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"46e7a275-1a83-4ee4-bbee-819afd9e1df6","Type":"ContainerStarted","Data":"55d139908fc565872075917de4d0a292f7ea2528f96553b97e4dba033641718b"} Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.394155 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.3941359540000002 podStartE2EDuration="2.394135954s" podCreationTimestamp="2025-10-04 04:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:41:05.39320727 +0000 UTC m=+5876.685216982" watchObservedRunningTime="2025-10-04 04:41:05.394135954 +0000 UTC m=+5876.686145666" Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.421878 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.421855708 podStartE2EDuration="2.421855708s" podCreationTimestamp="2025-10-04 04:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:41:05.419763624 +0000 UTC m=+5876.711773336" watchObservedRunningTime="2025-10-04 04:41:05.421855708 +0000 UTC m=+5876.713865420" Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.436322 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.436301237 podStartE2EDuration="2.436301237s" podCreationTimestamp="2025-10-04 04:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:41:05.43415974 +0000 UTC m=+5876.726169452" watchObservedRunningTime="2025-10-04 04:41:05.436301237 +0000 UTC m=+5876.728310969" Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.472922 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.472900663 podStartE2EDuration="2.472900663s" podCreationTimestamp="2025-10-04 04:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:41:05.466963737 +0000 UTC m=+5876.758973449" watchObservedRunningTime="2025-10-04 04:41:05.472900663 +0000 UTC m=+5876.764910375" Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.507429 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.507403624 podStartE2EDuration="2.507403624s" podCreationTimestamp="2025-10-04 04:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:41:05.489785314 +0000 UTC m=+5876.781795026" watchObservedRunningTime="2025-10-04 04:41:05.507403624 +0000 UTC m=+5876.799413336" Oct 04 04:41:05 crc kubenswrapper[4770]: I1004 04:41:05.683344 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39e02f97-9bb3-43b4-a2f8-7633252519a3" path="/var/lib/kubelet/pods/39e02f97-9bb3-43b4-a2f8-7633252519a3/volumes" Oct 04 04:41:09 crc kubenswrapper[4770]: I1004 04:41:09.086192 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:41:09 crc kubenswrapper[4770]: I1004 04:41:09.086934 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:41:09 crc kubenswrapper[4770]: I1004 04:41:09.109116 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:41:09 crc kubenswrapper[4770]: I1004 04:41:09.126667 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.363735 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=7.36371546 podStartE2EDuration="7.36371546s" podCreationTimestamp="2025-10-04 04:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:41:05.515558867 +0000 UTC m=+5876.807568589" watchObservedRunningTime="2025-10-04 04:41:10.36371546 +0000 UTC m=+5881.655725172" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.368530 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.370001 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.371666 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.382722 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.533970 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b502371-11cc-4e4e-94bb-8f97a61edcec-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.534065 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.534089 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.534152 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-scripts\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.534185 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.534220 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qk4b\" (UniqueName: \"kubernetes.io/projected/6b502371-11cc-4e4e-94bb-8f97a61edcec-kube-api-access-4qk4b\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.635597 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b502371-11cc-4e4e-94bb-8f97a61edcec-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.635669 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.635693 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.635736 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b502371-11cc-4e4e-94bb-8f97a61edcec-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.635765 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-scripts\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.635899 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.635963 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qk4b\" (UniqueName: \"kubernetes.io/projected/6b502371-11cc-4e4e-94bb-8f97a61edcec-kube-api-access-4qk4b\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.641890 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-scripts\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.642152 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.642494 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.643212 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.660917 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qk4b\" (UniqueName: \"kubernetes.io/projected/6b502371-11cc-4e4e-94bb-8f97a61edcec-kube-api-access-4qk4b\") pod \"cinder-scheduler-0\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:10 crc kubenswrapper[4770]: I1004 04:41:10.693129 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:41:11 crc kubenswrapper[4770]: I1004 04:41:11.140503 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:41:11 crc kubenswrapper[4770]: I1004 04:41:11.428598 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b502371-11cc-4e4e-94bb-8f97a61edcec","Type":"ContainerStarted","Data":"6cb98ab4ceb9273651ee39534df5c695747f5b1be856ccbea9ff5222613d73b8"} Oct 04 04:41:11 crc kubenswrapper[4770]: I1004 04:41:11.900750 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:41:11 crc kubenswrapper[4770]: I1004 04:41:11.901067 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="99081418-459d-4d41-9d13-447f409f1d8f" containerName="cinder-api-log" containerID="cri-o://bfe37593f19384cca39e89138fb81f3d89bccb96202c8b8910482b4d51e07959" gracePeriod=30 Oct 04 04:41:11 crc kubenswrapper[4770]: I1004 04:41:11.901241 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="99081418-459d-4d41-9d13-447f409f1d8f" containerName="cinder-api" containerID="cri-o://ee9ef5ebc010bc349dc626a5f4a465ae9408be0d70fdb133d406dba8ec69bb1a" gracePeriod=30 Oct 04 04:41:12 crc kubenswrapper[4770]: I1004 04:41:12.462729 4770 generic.go:334] "Generic (PLEG): container finished" podID="99081418-459d-4d41-9d13-447f409f1d8f" containerID="bfe37593f19384cca39e89138fb81f3d89bccb96202c8b8910482b4d51e07959" exitCode=143 Oct 04 04:41:12 crc kubenswrapper[4770]: I1004 04:41:12.462800 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99081418-459d-4d41-9d13-447f409f1d8f","Type":"ContainerDied","Data":"bfe37593f19384cca39e89138fb81f3d89bccb96202c8b8910482b4d51e07959"} Oct 04 04:41:12 crc kubenswrapper[4770]: I1004 04:41:12.474328 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b502371-11cc-4e4e-94bb-8f97a61edcec","Type":"ContainerStarted","Data":"37f892d29a3630bcc245e166b16fab749a21c108678cd7f49053838ecc4622ab"} Oct 04 04:41:12 crc kubenswrapper[4770]: I1004 04:41:12.976830 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 04 04:41:12 crc kubenswrapper[4770]: I1004 04:41:12.979434 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:12 crc kubenswrapper[4770]: I1004 04:41:12.991209 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.025447 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.086887 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.086948 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-dev\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087125 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/328aff13-c574-4896-b312-cb36ac8315c2-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087160 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087198 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087251 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087328 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087361 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087445 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-sys\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087529 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087600 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087629 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087667 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087692 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr9vf\" (UniqueName: \"kubernetes.io/projected/328aff13-c574-4896-b312-cb36ac8315c2-kube-api-access-qr9vf\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087733 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.087763 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-run\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189688 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189746 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189773 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189793 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr9vf\" (UniqueName: \"kubernetes.io/projected/328aff13-c574-4896-b312-cb36ac8315c2-kube-api-access-qr9vf\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189828 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189862 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-run\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189855 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189919 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189904 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189879 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.189969 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-dev\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190001 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/328aff13-c574-4896-b312-cb36ac8315c2-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190039 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190060 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190089 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190114 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190134 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190168 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-sys\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190200 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190506 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190554 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190862 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-dev\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190934 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.190975 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-run\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.191042 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-sys\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.191606 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/328aff13-c574-4896-b312-cb36ac8315c2-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.196914 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.197946 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/328aff13-c574-4896-b312-cb36ac8315c2-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.198718 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.206508 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.207128 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/328aff13-c574-4896-b312-cb36ac8315c2-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.218144 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr9vf\" (UniqueName: \"kubernetes.io/projected/328aff13-c574-4896-b312-cb36ac8315c2-kube-api-access-qr9vf\") pod \"cinder-volume-volume1-0\" (UID: \"328aff13-c574-4896-b312-cb36ac8315c2\") " pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.325558 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.486697 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b502371-11cc-4e4e-94bb-8f97a61edcec","Type":"ContainerStarted","Data":"73f3200e390ba3ced0801229f50cb5f43ad6f9d51ce6c6c909675aa274c175ae"} Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.519169 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.51915184 podStartE2EDuration="3.51915184s" podCreationTimestamp="2025-10-04 04:41:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:41:13.51453586 +0000 UTC m=+5884.806545572" watchObservedRunningTime="2025-10-04 04:41:13.51915184 +0000 UTC m=+5884.811161552" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.663795 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.665350 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.671267 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.707956 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.708037 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.805694 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-scripts\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.805744 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.805772 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5cbf3298-0dbe-4979-a94f-a6baabc7b545-ceph\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.805790 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.805900 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.805954 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-dev\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806084 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-sys\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806139 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-lib-modules\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806169 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806211 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-config-data\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806251 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806278 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806388 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806429 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjr86\" (UniqueName: \"kubernetes.io/projected/5cbf3298-0dbe-4979-a94f-a6baabc7b545-kube-api-access-mjr86\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806448 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.806857 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-run\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.825031 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.825081 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.886413 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908732 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908815 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjr86\" (UniqueName: \"kubernetes.io/projected/5cbf3298-0dbe-4979-a94f-a6baabc7b545-kube-api-access-mjr86\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908838 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908867 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-run\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908901 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908931 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-scripts\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908949 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-run\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908952 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908986 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5cbf3298-0dbe-4979-a94f-a6baabc7b545-ceph\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908994 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.908867 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909072 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909036 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909168 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909209 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-dev\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909237 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-sys\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909283 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-lib-modules\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909314 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909316 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-dev\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909368 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909392 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-sys\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909377 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909407 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-config-data\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909419 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5cbf3298-0dbe-4979-a94f-a6baabc7b545-lib-modules\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909501 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.909542 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.919809 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5cbf3298-0dbe-4979-a94f-a6baabc7b545-ceph\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.920105 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.920466 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-scripts\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.922531 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.923780 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbf3298-0dbe-4979-a94f-a6baabc7b545-config-data\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.927488 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 04:41:13 crc kubenswrapper[4770]: I1004 04:41:13.933501 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjr86\" (UniqueName: \"kubernetes.io/projected/5cbf3298-0dbe-4979-a94f-a6baabc7b545-kube-api-access-mjr86\") pod \"cinder-backup-0\" (UID: \"5cbf3298-0dbe-4979-a94f-a6baabc7b545\") " pod="openstack/cinder-backup-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.011671 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.085972 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.086074 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.109520 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.126534 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.144205 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.162910 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.495911 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"328aff13-c574-4896-b312-cb36ac8315c2","Type":"ContainerStarted","Data":"f20720eca6103b53ae48b884881b4c897dfcdadbf8010fad4512170528f7c38e"} Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.509659 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.548760 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.620951 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 04 04:41:14 crc kubenswrapper[4770]: W1004 04:41:14.643530 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbf3298_0dbe_4979_a94f_a6baabc7b545.slice/crio-e6876614a8dfe1548e8ed515cd6e8953a013a404e85428eaca7912c0565150c3 WatchSource:0}: Error finding container e6876614a8dfe1548e8ed515cd6e8953a013a404e85428eaca7912c0565150c3: Status 404 returned error can't find the container with id e6876614a8dfe1548e8ed515cd6e8953a013a404e85428eaca7912c0565150c3 Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.865221 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.82:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:41:14 crc kubenswrapper[4770]: I1004 04:41:14.906385 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.82:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:41:15 crc kubenswrapper[4770]: I1004 04:41:15.169207 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:41:15 crc kubenswrapper[4770]: I1004 04:41:15.169207 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:41:15 crc kubenswrapper[4770]: I1004 04:41:15.508870 4770 generic.go:334] "Generic (PLEG): container finished" podID="99081418-459d-4d41-9d13-447f409f1d8f" containerID="ee9ef5ebc010bc349dc626a5f4a465ae9408be0d70fdb133d406dba8ec69bb1a" exitCode=0 Oct 04 04:41:15 crc kubenswrapper[4770]: I1004 04:41:15.508964 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99081418-459d-4d41-9d13-447f409f1d8f","Type":"ContainerDied","Data":"ee9ef5ebc010bc349dc626a5f4a465ae9408be0d70fdb133d406dba8ec69bb1a"} Oct 04 04:41:15 crc kubenswrapper[4770]: I1004 04:41:15.512897 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5cbf3298-0dbe-4979-a94f-a6baabc7b545","Type":"ContainerStarted","Data":"e6876614a8dfe1548e8ed515cd6e8953a013a404e85428eaca7912c0565150c3"} Oct 04 04:41:15 crc kubenswrapper[4770]: I1004 04:41:15.700379 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 04:41:15 crc kubenswrapper[4770]: I1004 04:41:15.909613 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.063813 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klncv\" (UniqueName: \"kubernetes.io/projected/99081418-459d-4d41-9d13-447f409f1d8f-kube-api-access-klncv\") pod \"99081418-459d-4d41-9d13-447f409f1d8f\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.063881 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data\") pod \"99081418-459d-4d41-9d13-447f409f1d8f\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.064089 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-scripts\") pod \"99081418-459d-4d41-9d13-447f409f1d8f\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.064177 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99081418-459d-4d41-9d13-447f409f1d8f-logs\") pod \"99081418-459d-4d41-9d13-447f409f1d8f\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.064245 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-combined-ca-bundle\") pod \"99081418-459d-4d41-9d13-447f409f1d8f\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.064281 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99081418-459d-4d41-9d13-447f409f1d8f-etc-machine-id\") pod \"99081418-459d-4d41-9d13-447f409f1d8f\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.064303 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data-custom\") pod \"99081418-459d-4d41-9d13-447f409f1d8f\" (UID: \"99081418-459d-4d41-9d13-447f409f1d8f\") " Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.064615 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99081418-459d-4d41-9d13-447f409f1d8f-logs" (OuterVolumeSpecName: "logs") pod "99081418-459d-4d41-9d13-447f409f1d8f" (UID: "99081418-459d-4d41-9d13-447f409f1d8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.064664 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99081418-459d-4d41-9d13-447f409f1d8f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "99081418-459d-4d41-9d13-447f409f1d8f" (UID: "99081418-459d-4d41-9d13-447f409f1d8f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.064984 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99081418-459d-4d41-9d13-447f409f1d8f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.065005 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99081418-459d-4d41-9d13-447f409f1d8f-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.071121 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-scripts" (OuterVolumeSpecName: "scripts") pod "99081418-459d-4d41-9d13-447f409f1d8f" (UID: "99081418-459d-4d41-9d13-447f409f1d8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.072189 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99081418-459d-4d41-9d13-447f409f1d8f-kube-api-access-klncv" (OuterVolumeSpecName: "kube-api-access-klncv") pod "99081418-459d-4d41-9d13-447f409f1d8f" (UID: "99081418-459d-4d41-9d13-447f409f1d8f"). InnerVolumeSpecName "kube-api-access-klncv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.072666 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "99081418-459d-4d41-9d13-447f409f1d8f" (UID: "99081418-459d-4d41-9d13-447f409f1d8f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.118689 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99081418-459d-4d41-9d13-447f409f1d8f" (UID: "99081418-459d-4d41-9d13-447f409f1d8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.156173 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data" (OuterVolumeSpecName: "config-data") pod "99081418-459d-4d41-9d13-447f409f1d8f" (UID: "99081418-459d-4d41-9d13-447f409f1d8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.166591 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.166641 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.166656 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.166668 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klncv\" (UniqueName: \"kubernetes.io/projected/99081418-459d-4d41-9d13-447f409f1d8f-kube-api-access-klncv\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.166679 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99081418-459d-4d41-9d13-447f409f1d8f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.529192 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99081418-459d-4d41-9d13-447f409f1d8f","Type":"ContainerDied","Data":"449f5de8def159c36f2529874350c4fa2ee87a74052313f0b79fa2e8bb9e65ac"} Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.529630 4770 scope.go:117] "RemoveContainer" containerID="ee9ef5ebc010bc349dc626a5f4a465ae9408be0d70fdb133d406dba8ec69bb1a" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.529238 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.539434 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"328aff13-c574-4896-b312-cb36ac8315c2","Type":"ContainerStarted","Data":"4475b5fcada50f4213e5be8963d68641560cef0360d4c6c4c6481846251f3953"} Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.645572 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.653078 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.662308 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:41:16 crc kubenswrapper[4770]: E1004 04:41:16.663105 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99081418-459d-4d41-9d13-447f409f1d8f" containerName="cinder-api" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.663125 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="99081418-459d-4d41-9d13-447f409f1d8f" containerName="cinder-api" Oct 04 04:41:16 crc kubenswrapper[4770]: E1004 04:41:16.663168 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99081418-459d-4d41-9d13-447f409f1d8f" containerName="cinder-api-log" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.663176 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="99081418-459d-4d41-9d13-447f409f1d8f" containerName="cinder-api-log" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.663366 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="99081418-459d-4d41-9d13-447f409f1d8f" containerName="cinder-api" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.663411 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="99081418-459d-4d41-9d13-447f409f1d8f" containerName="cinder-api-log" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.664604 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.667937 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.670177 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.730536 4770 scope.go:117] "RemoveContainer" containerID="bfe37593f19384cca39e89138fb81f3d89bccb96202c8b8910482b4d51e07959" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.808725 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-config-data-custom\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.808782 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.808803 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-config-data\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.808823 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3f8f6e6-6f18-40fc-842b-73323e55e559-logs\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.808861 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-scripts\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.808887 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c3f8f6e6-6f18-40fc-842b-73323e55e559-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.808957 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr9v2\" (UniqueName: \"kubernetes.io/projected/c3f8f6e6-6f18-40fc-842b-73323e55e559-kube-api-access-tr9v2\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.910212 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr9v2\" (UniqueName: \"kubernetes.io/projected/c3f8f6e6-6f18-40fc-842b-73323e55e559-kube-api-access-tr9v2\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.910537 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-config-data-custom\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.910630 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.910714 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-config-data\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.910782 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3f8f6e6-6f18-40fc-842b-73323e55e559-logs\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.910869 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-scripts\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.910946 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c3f8f6e6-6f18-40fc-842b-73323e55e559-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.911166 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c3f8f6e6-6f18-40fc-842b-73323e55e559-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.912392 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3f8f6e6-6f18-40fc-842b-73323e55e559-logs\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.917698 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.917714 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-config-data\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.918395 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-config-data-custom\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.923289 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3f8f6e6-6f18-40fc-842b-73323e55e559-scripts\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.937937 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr9v2\" (UniqueName: \"kubernetes.io/projected/c3f8f6e6-6f18-40fc-842b-73323e55e559-kube-api-access-tr9v2\") pod \"cinder-api-0\" (UID: \"c3f8f6e6-6f18-40fc-842b-73323e55e559\") " pod="openstack/cinder-api-0" Oct 04 04:41:16 crc kubenswrapper[4770]: I1004 04:41:16.998388 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:41:17 crc kubenswrapper[4770]: W1004 04:41:17.506598 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3f8f6e6_6f18_40fc_842b_73323e55e559.slice/crio-2912679d6e3fa2f78ae814f85b25b14edc35fb8123e78249c1c5592ba5ae3029 WatchSource:0}: Error finding container 2912679d6e3fa2f78ae814f85b25b14edc35fb8123e78249c1c5592ba5ae3029: Status 404 returned error can't find the container with id 2912679d6e3fa2f78ae814f85b25b14edc35fb8123e78249c1c5592ba5ae3029 Oct 04 04:41:17 crc kubenswrapper[4770]: I1004 04:41:17.507910 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:41:17 crc kubenswrapper[4770]: I1004 04:41:17.550269 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5cbf3298-0dbe-4979-a94f-a6baabc7b545","Type":"ContainerStarted","Data":"c55db3183fb3900919f9fe48d2e588dde34ff70f06e2310ae39c255d6d9343fa"} Oct 04 04:41:17 crc kubenswrapper[4770]: I1004 04:41:17.550312 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5cbf3298-0dbe-4979-a94f-a6baabc7b545","Type":"ContainerStarted","Data":"b6375913425d1688b57cb6b5af87c78557abf8eb4ab653367f5afbe289716ea9"} Oct 04 04:41:17 crc kubenswrapper[4770]: I1004 04:41:17.552870 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"328aff13-c574-4896-b312-cb36ac8315c2","Type":"ContainerStarted","Data":"4e03fc5f1839932c34c496fde9061a7e0a8e95716df05768ba4ef2ee46657f51"} Oct 04 04:41:17 crc kubenswrapper[4770]: I1004 04:41:17.554267 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c3f8f6e6-6f18-40fc-842b-73323e55e559","Type":"ContainerStarted","Data":"2912679d6e3fa2f78ae814f85b25b14edc35fb8123e78249c1c5592ba5ae3029"} Oct 04 04:41:17 crc kubenswrapper[4770]: I1004 04:41:17.685692 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99081418-459d-4d41-9d13-447f409f1d8f" path="/var/lib/kubelet/pods/99081418-459d-4d41-9d13-447f409f1d8f/volumes" Oct 04 04:41:18 crc kubenswrapper[4770]: I1004 04:41:18.326510 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:18 crc kubenswrapper[4770]: I1004 04:41:18.569650 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c3f8f6e6-6f18-40fc-842b-73323e55e559","Type":"ContainerStarted","Data":"e720482daf6d0551b5850cbfb6e25439ddae544f601b582b8f7030b4f7e3d429"} Oct 04 04:41:18 crc kubenswrapper[4770]: I1004 04:41:18.606629 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.7037593829999995 podStartE2EDuration="6.606608536s" podCreationTimestamp="2025-10-04 04:41:12 +0000 UTC" firstStartedPulling="2025-10-04 04:41:13.895089456 +0000 UTC m=+5885.187099168" lastFinishedPulling="2025-10-04 04:41:15.797938609 +0000 UTC m=+5887.089948321" observedRunningTime="2025-10-04 04:41:17.587371227 +0000 UTC m=+5888.879380959" watchObservedRunningTime="2025-10-04 04:41:18.606608536 +0000 UTC m=+5889.898618258" Oct 04 04:41:18 crc kubenswrapper[4770]: I1004 04:41:18.611757 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.517555258 podStartE2EDuration="5.611744521s" podCreationTimestamp="2025-10-04 04:41:13 +0000 UTC" firstStartedPulling="2025-10-04 04:41:14.646455624 +0000 UTC m=+5885.938465336" lastFinishedPulling="2025-10-04 04:41:16.740644887 +0000 UTC m=+5888.032654599" observedRunningTime="2025-10-04 04:41:18.602090538 +0000 UTC m=+5889.894100270" watchObservedRunningTime="2025-10-04 04:41:18.611744521 +0000 UTC m=+5889.903754233" Oct 04 04:41:19 crc kubenswrapper[4770]: I1004 04:41:19.012405 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 04 04:41:19 crc kubenswrapper[4770]: I1004 04:41:19.585822 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c3f8f6e6-6f18-40fc-842b-73323e55e559","Type":"ContainerStarted","Data":"9f943e201d376a5896e4e9819ed9dc17713c1c700f65ec6ce4d8888176c5d576"} Oct 04 04:41:19 crc kubenswrapper[4770]: I1004 04:41:19.624036 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.623992747 podStartE2EDuration="3.623992747s" podCreationTimestamp="2025-10-04 04:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:41:19.612204728 +0000 UTC m=+5890.904214440" watchObservedRunningTime="2025-10-04 04:41:19.623992747 +0000 UTC m=+5890.916002479" Oct 04 04:41:20 crc kubenswrapper[4770]: I1004 04:41:20.599089 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 04:41:20 crc kubenswrapper[4770]: I1004 04:41:20.912901 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 04:41:20 crc kubenswrapper[4770]: I1004 04:41:20.971277 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:41:21 crc kubenswrapper[4770]: I1004 04:41:21.606794 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerName="cinder-scheduler" containerID="cri-o://37f892d29a3630bcc245e166b16fab749a21c108678cd7f49053838ecc4622ab" gracePeriod=30 Oct 04 04:41:21 crc kubenswrapper[4770]: I1004 04:41:21.607663 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerName="probe" containerID="cri-o://73f3200e390ba3ced0801229f50cb5f43ad6f9d51ce6c6c909675aa274c175ae" gracePeriod=30 Oct 04 04:41:22 crc kubenswrapper[4770]: I1004 04:41:22.618965 4770 generic.go:334] "Generic (PLEG): container finished" podID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerID="73f3200e390ba3ced0801229f50cb5f43ad6f9d51ce6c6c909675aa274c175ae" exitCode=0 Oct 04 04:41:22 crc kubenswrapper[4770]: I1004 04:41:22.619326 4770 generic.go:334] "Generic (PLEG): container finished" podID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerID="37f892d29a3630bcc245e166b16fab749a21c108678cd7f49053838ecc4622ab" exitCode=0 Oct 04 04:41:22 crc kubenswrapper[4770]: I1004 04:41:22.619036 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b502371-11cc-4e4e-94bb-8f97a61edcec","Type":"ContainerDied","Data":"73f3200e390ba3ced0801229f50cb5f43ad6f9d51ce6c6c909675aa274c175ae"} Oct 04 04:41:22 crc kubenswrapper[4770]: I1004 04:41:22.619369 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b502371-11cc-4e4e-94bb-8f97a61edcec","Type":"ContainerDied","Data":"37f892d29a3630bcc245e166b16fab749a21c108678cd7f49053838ecc4622ab"} Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.192039 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.376875 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data\") pod \"6b502371-11cc-4e4e-94bb-8f97a61edcec\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.376992 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qk4b\" (UniqueName: \"kubernetes.io/projected/6b502371-11cc-4e4e-94bb-8f97a61edcec-kube-api-access-4qk4b\") pod \"6b502371-11cc-4e4e-94bb-8f97a61edcec\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.377123 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-scripts\") pod \"6b502371-11cc-4e4e-94bb-8f97a61edcec\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.377217 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-combined-ca-bundle\") pod \"6b502371-11cc-4e4e-94bb-8f97a61edcec\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.377266 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data-custom\") pod \"6b502371-11cc-4e4e-94bb-8f97a61edcec\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.377308 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b502371-11cc-4e4e-94bb-8f97a61edcec-etc-machine-id\") pod \"6b502371-11cc-4e4e-94bb-8f97a61edcec\" (UID: \"6b502371-11cc-4e4e-94bb-8f97a61edcec\") " Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.377775 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b502371-11cc-4e4e-94bb-8f97a61edcec-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6b502371-11cc-4e4e-94bb-8f97a61edcec" (UID: "6b502371-11cc-4e4e-94bb-8f97a61edcec"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.382425 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6b502371-11cc-4e4e-94bb-8f97a61edcec" (UID: "6b502371-11cc-4e4e-94bb-8f97a61edcec"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.395313 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b502371-11cc-4e4e-94bb-8f97a61edcec-kube-api-access-4qk4b" (OuterVolumeSpecName: "kube-api-access-4qk4b") pod "6b502371-11cc-4e4e-94bb-8f97a61edcec" (UID: "6b502371-11cc-4e4e-94bb-8f97a61edcec"). InnerVolumeSpecName "kube-api-access-4qk4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.400022 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-scripts" (OuterVolumeSpecName: "scripts") pod "6b502371-11cc-4e4e-94bb-8f97a61edcec" (UID: "6b502371-11cc-4e4e-94bb-8f97a61edcec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.443818 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b502371-11cc-4e4e-94bb-8f97a61edcec" (UID: "6b502371-11cc-4e4e-94bb-8f97a61edcec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.477073 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data" (OuterVolumeSpecName: "config-data") pod "6b502371-11cc-4e4e-94bb-8f97a61edcec" (UID: "6b502371-11cc-4e4e-94bb-8f97a61edcec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.479520 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.479551 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.479562 4770 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b502371-11cc-4e4e-94bb-8f97a61edcec-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.479572 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.479583 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qk4b\" (UniqueName: \"kubernetes.io/projected/6b502371-11cc-4e4e-94bb-8f97a61edcec-kube-api-access-4qk4b\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.479594 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b502371-11cc-4e4e-94bb-8f97a61edcec-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.567779 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.632677 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b502371-11cc-4e4e-94bb-8f97a61edcec","Type":"ContainerDied","Data":"6cb98ab4ceb9273651ee39534df5c695747f5b1be856ccbea9ff5222613d73b8"} Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.632727 4770 scope.go:117] "RemoveContainer" containerID="73f3200e390ba3ced0801229f50cb5f43ad6f9d51ce6c6c909675aa274c175ae" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.632840 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.665159 4770 scope.go:117] "RemoveContainer" containerID="37f892d29a3630bcc245e166b16fab749a21c108678cd7f49053838ecc4622ab" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.706294 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.706328 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.719276 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:41:23 crc kubenswrapper[4770]: E1004 04:41:23.719914 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerName="probe" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.719934 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerName="probe" Oct 04 04:41:23 crc kubenswrapper[4770]: E1004 04:41:23.719962 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerName="cinder-scheduler" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.719969 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerName="cinder-scheduler" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.720167 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerName="cinder-scheduler" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.720197 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b502371-11cc-4e4e-94bb-8f97a61edcec" containerName="probe" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.721173 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.723446 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.727971 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.834503 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.834559 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.835497 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.835537 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.838687 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.842668 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.885794 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5zqh\" (UniqueName: \"kubernetes.io/projected/3a2e4f90-e9b4-44ed-abef-2de3454b9260-kube-api-access-v5zqh\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.885874 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2e4f90-e9b4-44ed-abef-2de3454b9260-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.885906 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.885932 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.885955 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.886049 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.988360 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5zqh\" (UniqueName: \"kubernetes.io/projected/3a2e4f90-e9b4-44ed-abef-2de3454b9260-kube-api-access-v5zqh\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.988418 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2e4f90-e9b4-44ed-abef-2de3454b9260-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.988466 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.988508 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.988512 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a2e4f90-e9b4-44ed-abef-2de3454b9260-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.989179 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.989315 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.994823 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.995614 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:23 crc kubenswrapper[4770]: I1004 04:41:23.999516 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.001157 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a2e4f90-e9b4-44ed-abef-2de3454b9260-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.007923 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5zqh\" (UniqueName: \"kubernetes.io/projected/3a2e4f90-e9b4-44ed-abef-2de3454b9260-kube-api-access-v5zqh\") pod \"cinder-scheduler-0\" (UID: \"3a2e4f90-e9b4-44ed-abef-2de3454b9260\") " pod="openstack/cinder-scheduler-0" Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.047967 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.088832 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.089660 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.092598 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.095204 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.333469 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.598252 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:41:24 crc kubenswrapper[4770]: I1004 04:41:24.661990 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a2e4f90-e9b4-44ed-abef-2de3454b9260","Type":"ContainerStarted","Data":"e191568846c2448befa249fb7b3727fcef06bdcb2a44900b209762f5e56eb782"} Oct 04 04:41:25 crc kubenswrapper[4770]: I1004 04:41:25.706380 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b502371-11cc-4e4e-94bb-8f97a61edcec" path="/var/lib/kubelet/pods/6b502371-11cc-4e4e-94bb-8f97a61edcec/volumes" Oct 04 04:41:25 crc kubenswrapper[4770]: I1004 04:41:25.707601 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a2e4f90-e9b4-44ed-abef-2de3454b9260","Type":"ContainerStarted","Data":"ca3f2abf781012e0f92dbb4e8bba7f6d3fce1c619b72d484fc1fadbd1e4b5774"} Oct 04 04:41:25 crc kubenswrapper[4770]: I1004 04:41:25.807837 4770 scope.go:117] "RemoveContainer" containerID="f8ac077e3dcd2450d464e3c24e55a86c463b22fe37fd2483ae6a96cfc74f11a0" Oct 04 04:41:26 crc kubenswrapper[4770]: I1004 04:41:26.713755 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a2e4f90-e9b4-44ed-abef-2de3454b9260","Type":"ContainerStarted","Data":"96d421bd67e39932e2aac5ccaf60a60b2465c144d1c5dea1ee0c4d858d93ec8b"} Oct 04 04:41:26 crc kubenswrapper[4770]: I1004 04:41:26.740476 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.740452512 podStartE2EDuration="3.740452512s" podCreationTimestamp="2025-10-04 04:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:41:26.7326877 +0000 UTC m=+5898.024697422" watchObservedRunningTime="2025-10-04 04:41:26.740452512 +0000 UTC m=+5898.032462234" Oct 04 04:41:28 crc kubenswrapper[4770]: I1004 04:41:28.967191 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 04:41:29 crc kubenswrapper[4770]: I1004 04:41:29.049233 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 04:41:34 crc kubenswrapper[4770]: I1004 04:41:34.251510 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 04:42:41 crc kubenswrapper[4770]: I1004 04:42:41.830950 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gsz6p"] Oct 04 04:42:41 crc kubenswrapper[4770]: I1004 04:42:41.834405 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:41 crc kubenswrapper[4770]: I1004 04:42:41.864775 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gsz6p"] Oct 04 04:42:41 crc kubenswrapper[4770]: I1004 04:42:41.964677 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-utilities\") pod \"redhat-operators-gsz6p\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:41 crc kubenswrapper[4770]: I1004 04:42:41.965098 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29npl\" (UniqueName: \"kubernetes.io/projected/ffc21acb-6709-44f1-9684-b6e1055f5036-kube-api-access-29npl\") pod \"redhat-operators-gsz6p\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:41 crc kubenswrapper[4770]: I1004 04:42:41.965160 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-catalog-content\") pod \"redhat-operators-gsz6p\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:42 crc kubenswrapper[4770]: I1004 04:42:42.067083 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-utilities\") pod \"redhat-operators-gsz6p\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:42 crc kubenswrapper[4770]: I1004 04:42:42.067154 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29npl\" (UniqueName: \"kubernetes.io/projected/ffc21acb-6709-44f1-9684-b6e1055f5036-kube-api-access-29npl\") pod \"redhat-operators-gsz6p\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:42 crc kubenswrapper[4770]: I1004 04:42:42.067217 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-catalog-content\") pod \"redhat-operators-gsz6p\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:42 crc kubenswrapper[4770]: I1004 04:42:42.067620 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-utilities\") pod \"redhat-operators-gsz6p\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:42 crc kubenswrapper[4770]: I1004 04:42:42.067669 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-catalog-content\") pod \"redhat-operators-gsz6p\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:42 crc kubenswrapper[4770]: I1004 04:42:42.089939 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29npl\" (UniqueName: \"kubernetes.io/projected/ffc21acb-6709-44f1-9684-b6e1055f5036-kube-api-access-29npl\") pod \"redhat-operators-gsz6p\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:42 crc kubenswrapper[4770]: I1004 04:42:42.169230 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:42 crc kubenswrapper[4770]: I1004 04:42:42.630045 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gsz6p"] Oct 04 04:42:43 crc kubenswrapper[4770]: I1004 04:42:43.474371 4770 generic.go:334] "Generic (PLEG): container finished" podID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerID="42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819" exitCode=0 Oct 04 04:42:43 crc kubenswrapper[4770]: I1004 04:42:43.474437 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsz6p" event={"ID":"ffc21acb-6709-44f1-9684-b6e1055f5036","Type":"ContainerDied","Data":"42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819"} Oct 04 04:42:43 crc kubenswrapper[4770]: I1004 04:42:43.475791 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsz6p" event={"ID":"ffc21acb-6709-44f1-9684-b6e1055f5036","Type":"ContainerStarted","Data":"ce47fc20448456c98fb0815aff9e497175804310606fe2e33554662ecdbf1581"} Oct 04 04:42:43 crc kubenswrapper[4770]: I1004 04:42:43.478225 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:42:45 crc kubenswrapper[4770]: I1004 04:42:45.508775 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsz6p" event={"ID":"ffc21acb-6709-44f1-9684-b6e1055f5036","Type":"ContainerStarted","Data":"9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41"} Oct 04 04:42:46 crc kubenswrapper[4770]: I1004 04:42:46.520111 4770 generic.go:334] "Generic (PLEG): container finished" podID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerID="9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41" exitCode=0 Oct 04 04:42:46 crc kubenswrapper[4770]: I1004 04:42:46.520193 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsz6p" event={"ID":"ffc21acb-6709-44f1-9684-b6e1055f5036","Type":"ContainerDied","Data":"9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41"} Oct 04 04:42:48 crc kubenswrapper[4770]: I1004 04:42:48.547742 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsz6p" event={"ID":"ffc21acb-6709-44f1-9684-b6e1055f5036","Type":"ContainerStarted","Data":"547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd"} Oct 04 04:42:48 crc kubenswrapper[4770]: I1004 04:42:48.598618 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gsz6p" podStartSLOduration=3.805464329 podStartE2EDuration="7.598596876s" podCreationTimestamp="2025-10-04 04:42:41 +0000 UTC" firstStartedPulling="2025-10-04 04:42:43.477506202 +0000 UTC m=+5974.769515954" lastFinishedPulling="2025-10-04 04:42:47.270638799 +0000 UTC m=+5978.562648501" observedRunningTime="2025-10-04 04:42:48.591717776 +0000 UTC m=+5979.883727498" watchObservedRunningTime="2025-10-04 04:42:48.598596876 +0000 UTC m=+5979.890606608" Oct 04 04:42:52 crc kubenswrapper[4770]: I1004 04:42:52.169581 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:52 crc kubenswrapper[4770]: I1004 04:42:52.171198 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:52 crc kubenswrapper[4770]: I1004 04:42:52.240113 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:52 crc kubenswrapper[4770]: I1004 04:42:52.653948 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:52 crc kubenswrapper[4770]: I1004 04:42:52.713654 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gsz6p"] Oct 04 04:42:54 crc kubenswrapper[4770]: I1004 04:42:54.618390 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gsz6p" podUID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerName="registry-server" containerID="cri-o://547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd" gracePeriod=2 Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.063545 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.109562 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-catalog-content\") pod \"ffc21acb-6709-44f1-9684-b6e1055f5036\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.109669 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-utilities\") pod \"ffc21acb-6709-44f1-9684-b6e1055f5036\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.109787 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29npl\" (UniqueName: \"kubernetes.io/projected/ffc21acb-6709-44f1-9684-b6e1055f5036-kube-api-access-29npl\") pod \"ffc21acb-6709-44f1-9684-b6e1055f5036\" (UID: \"ffc21acb-6709-44f1-9684-b6e1055f5036\") " Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.110998 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-utilities" (OuterVolumeSpecName: "utilities") pod "ffc21acb-6709-44f1-9684-b6e1055f5036" (UID: "ffc21acb-6709-44f1-9684-b6e1055f5036"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.116715 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc21acb-6709-44f1-9684-b6e1055f5036-kube-api-access-29npl" (OuterVolumeSpecName: "kube-api-access-29npl") pod "ffc21acb-6709-44f1-9684-b6e1055f5036" (UID: "ffc21acb-6709-44f1-9684-b6e1055f5036"). InnerVolumeSpecName "kube-api-access-29npl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.194735 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ffc21acb-6709-44f1-9684-b6e1055f5036" (UID: "ffc21acb-6709-44f1-9684-b6e1055f5036"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.213204 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.213269 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc21acb-6709-44f1-9684-b6e1055f5036-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.213284 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29npl\" (UniqueName: \"kubernetes.io/projected/ffc21acb-6709-44f1-9684-b6e1055f5036-kube-api-access-29npl\") on node \"crc\" DevicePath \"\"" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.632171 4770 generic.go:334] "Generic (PLEG): container finished" podID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerID="547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd" exitCode=0 Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.632206 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsz6p" event={"ID":"ffc21acb-6709-44f1-9684-b6e1055f5036","Type":"ContainerDied","Data":"547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd"} Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.632254 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gsz6p" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.633374 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gsz6p" event={"ID":"ffc21acb-6709-44f1-9684-b6e1055f5036","Type":"ContainerDied","Data":"ce47fc20448456c98fb0815aff9e497175804310606fe2e33554662ecdbf1581"} Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.633389 4770 scope.go:117] "RemoveContainer" containerID="547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.671305 4770 scope.go:117] "RemoveContainer" containerID="9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.692041 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gsz6p"] Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.692392 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gsz6p"] Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.696424 4770 scope.go:117] "RemoveContainer" containerID="42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.769647 4770 scope.go:117] "RemoveContainer" containerID="547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd" Oct 04 04:42:55 crc kubenswrapper[4770]: E1004 04:42:55.771472 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd\": container with ID starting with 547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd not found: ID does not exist" containerID="547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.771516 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd"} err="failed to get container status \"547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd\": rpc error: code = NotFound desc = could not find container \"547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd\": container with ID starting with 547d238ab13b28f41bcd6ba183a0e32fa87a2ef981cf141e913cb4e2222e68fd not found: ID does not exist" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.771541 4770 scope.go:117] "RemoveContainer" containerID="9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41" Oct 04 04:42:55 crc kubenswrapper[4770]: E1004 04:42:55.772624 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41\": container with ID starting with 9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41 not found: ID does not exist" containerID="9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.772696 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41"} err="failed to get container status \"9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41\": rpc error: code = NotFound desc = could not find container \"9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41\": container with ID starting with 9bdf64a1493639d92215b682224143d378ddcfe3f520f97ac668a219699b1b41 not found: ID does not exist" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.772747 4770 scope.go:117] "RemoveContainer" containerID="42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819" Oct 04 04:42:55 crc kubenswrapper[4770]: E1004 04:42:55.773463 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819\": container with ID starting with 42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819 not found: ID does not exist" containerID="42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819" Oct 04 04:42:55 crc kubenswrapper[4770]: I1004 04:42:55.773517 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819"} err="failed to get container status \"42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819\": rpc error: code = NotFound desc = could not find container \"42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819\": container with ID starting with 42f6ac17278bf4bc9e313c8696a4b0bf7c0cc1f530bd129a0de5330c62dce819 not found: ID does not exist" Oct 04 04:42:57 crc kubenswrapper[4770]: I1004 04:42:57.686485 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc21acb-6709-44f1-9684-b6e1055f5036" path="/var/lib/kubelet/pods/ffc21acb-6709-44f1-9684-b6e1055f5036/volumes" Oct 04 04:43:03 crc kubenswrapper[4770]: I1004 04:43:03.036645 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-z58nf"] Oct 04 04:43:03 crc kubenswrapper[4770]: I1004 04:43:03.045415 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-z58nf"] Oct 04 04:43:03 crc kubenswrapper[4770]: I1004 04:43:03.708516 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc0de4a6-c99b-46c8-95d5-713076a5ed08" path="/var/lib/kubelet/pods/bc0de4a6-c99b-46c8-95d5-713076a5ed08/volumes" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.674498 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-857xz"] Oct 04 04:43:12 crc kubenswrapper[4770]: E1004 04:43:12.675447 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerName="extract-content" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.675461 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerName="extract-content" Oct 04 04:43:12 crc kubenswrapper[4770]: E1004 04:43:12.675495 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerName="registry-server" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.675500 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerName="registry-server" Oct 04 04:43:12 crc kubenswrapper[4770]: E1004 04:43:12.675514 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerName="extract-utilities" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.675519 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerName="extract-utilities" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.675700 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc21acb-6709-44f1-9684-b6e1055f5036" containerName="registry-server" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.677181 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.714689 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-857xz"] Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.853989 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-utilities\") pod \"certified-operators-857xz\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.854179 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-catalog-content\") pod \"certified-operators-857xz\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.854213 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj7xv\" (UniqueName: \"kubernetes.io/projected/756a79db-b81a-4095-93a3-0c5ecbbaacf6-kube-api-access-dj7xv\") pod \"certified-operators-857xz\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.872517 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cnp42"] Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.874601 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.880928 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cnp42"] Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.955961 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-catalog-content\") pod \"certified-operators-857xz\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.956050 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj7xv\" (UniqueName: \"kubernetes.io/projected/756a79db-b81a-4095-93a3-0c5ecbbaacf6-kube-api-access-dj7xv\") pod \"certified-operators-857xz\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.956134 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-utilities\") pod \"certified-operators-857xz\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.956510 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-catalog-content\") pod \"certified-operators-857xz\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.956685 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-utilities\") pod \"certified-operators-857xz\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:12 crc kubenswrapper[4770]: I1004 04:43:12.979859 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj7xv\" (UniqueName: \"kubernetes.io/projected/756a79db-b81a-4095-93a3-0c5ecbbaacf6-kube-api-access-dj7xv\") pod \"certified-operators-857xz\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.005287 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.044861 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-eef1-account-create-6cp4q"] Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.055267 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-eef1-account-create-6cp4q"] Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.057979 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-utilities\") pod \"community-operators-cnp42\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.058242 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzrc6\" (UniqueName: \"kubernetes.io/projected/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-kube-api-access-dzrc6\") pod \"community-operators-cnp42\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.058268 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-catalog-content\") pod \"community-operators-cnp42\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.164794 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-utilities\") pod \"community-operators-cnp42\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.164977 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzrc6\" (UniqueName: \"kubernetes.io/projected/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-kube-api-access-dzrc6\") pod \"community-operators-cnp42\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.165023 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-catalog-content\") pod \"community-operators-cnp42\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.165395 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-utilities\") pod \"community-operators-cnp42\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.167954 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-catalog-content\") pod \"community-operators-cnp42\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.202373 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzrc6\" (UniqueName: \"kubernetes.io/projected/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-kube-api-access-dzrc6\") pod \"community-operators-cnp42\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.378201 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-857xz"] Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.491188 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.685638 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a97b61-8706-46df-847c-5c5a5e3528c2" path="/var/lib/kubelet/pods/42a97b61-8706-46df-847c-5c5a5e3528c2/volumes" Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.849504 4770 generic.go:334] "Generic (PLEG): container finished" podID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerID="86254b538369fdc5053b8cb9ca177a51d4117311c37a7c3d3ced2beafe0b0058" exitCode=0 Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.849549 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-857xz" event={"ID":"756a79db-b81a-4095-93a3-0c5ecbbaacf6","Type":"ContainerDied","Data":"86254b538369fdc5053b8cb9ca177a51d4117311c37a7c3d3ced2beafe0b0058"} Oct 04 04:43:13 crc kubenswrapper[4770]: I1004 04:43:13.849574 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-857xz" event={"ID":"756a79db-b81a-4095-93a3-0c5ecbbaacf6","Type":"ContainerStarted","Data":"6dae64bb9b321e25c502effac9cca7614beae1e19dc6c430d84a5b187f020769"} Oct 04 04:43:14 crc kubenswrapper[4770]: I1004 04:43:14.028486 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cnp42"] Oct 04 04:43:14 crc kubenswrapper[4770]: W1004 04:43:14.032665 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dab3d1c_4281_4d29_b9b2_f86e4b0e5a16.slice/crio-7c6422c2f761dcb29852714227d067ff81fd214091f50bf64e509c9124a5d427 WatchSource:0}: Error finding container 7c6422c2f761dcb29852714227d067ff81fd214091f50bf64e509c9124a5d427: Status 404 returned error can't find the container with id 7c6422c2f761dcb29852714227d067ff81fd214091f50bf64e509c9124a5d427 Oct 04 04:43:14 crc kubenswrapper[4770]: I1004 04:43:14.858409 4770 generic.go:334] "Generic (PLEG): container finished" podID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerID="d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615" exitCode=0 Oct 04 04:43:14 crc kubenswrapper[4770]: I1004 04:43:14.858486 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnp42" event={"ID":"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16","Type":"ContainerDied","Data":"d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615"} Oct 04 04:43:14 crc kubenswrapper[4770]: I1004 04:43:14.858761 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnp42" event={"ID":"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16","Type":"ContainerStarted","Data":"7c6422c2f761dcb29852714227d067ff81fd214091f50bf64e509c9124a5d427"} Oct 04 04:43:16 crc kubenswrapper[4770]: I1004 04:43:16.885484 4770 generic.go:334] "Generic (PLEG): container finished" podID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerID="d6fe3268d09a0b69fc11d30ef57314bfa20fba09100e52d7002f2a00fc56c449" exitCode=0 Oct 04 04:43:16 crc kubenswrapper[4770]: I1004 04:43:16.885913 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-857xz" event={"ID":"756a79db-b81a-4095-93a3-0c5ecbbaacf6","Type":"ContainerDied","Data":"d6fe3268d09a0b69fc11d30ef57314bfa20fba09100e52d7002f2a00fc56c449"} Oct 04 04:43:18 crc kubenswrapper[4770]: I1004 04:43:18.026354 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-jxs7r"] Oct 04 04:43:18 crc kubenswrapper[4770]: I1004 04:43:18.033993 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-jxs7r"] Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.204205 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ncsvk"] Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.206551 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.209795 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-k4k7v" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.212205 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.218929 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ncsvk"] Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.228303 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-rwzj2"] Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.230261 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.268993 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rwzj2"] Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.391124 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-scripts\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.391176 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-var-log\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.391374 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-var-lib\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.391424 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-var-log-ovn\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.391448 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l2f6\" (UniqueName: \"kubernetes.io/projected/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-kube-api-access-4l2f6\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.391579 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-var-run-ovn\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.391650 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-var-run\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.391903 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-etc-ovs\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.391985 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-var-run\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.392039 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz6c9\" (UniqueName: \"kubernetes.io/projected/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-kube-api-access-fz6c9\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.392077 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-scripts\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.493746 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-etc-ovs\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.493793 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-var-run\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.493812 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz6c9\" (UniqueName: \"kubernetes.io/projected/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-kube-api-access-fz6c9\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.493837 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-scripts\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.493860 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-scripts\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.493882 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-var-log\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.493933 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-var-lib\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.493953 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-var-log-ovn\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.493971 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l2f6\" (UniqueName: \"kubernetes.io/projected/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-kube-api-access-4l2f6\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.494065 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-var-run\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.494089 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-var-run-ovn\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.494159 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-etc-ovs\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.494187 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-var-log\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.494190 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-var-lib\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.494189 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-var-run\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.494244 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-var-run-ovn\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.494259 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-var-log-ovn\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.494278 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-var-run\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.496477 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-scripts\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.496713 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-scripts\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.516369 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l2f6\" (UniqueName: \"kubernetes.io/projected/8e6ae49d-7530-44a5-bfb1-6201c9e745c8-kube-api-access-4l2f6\") pod \"ovn-controller-ncsvk\" (UID: \"8e6ae49d-7530-44a5-bfb1-6201c9e745c8\") " pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.520705 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz6c9\" (UniqueName: \"kubernetes.io/projected/ea243e32-f3d6-47a6-b2fe-9728af6bc06a-kube-api-access-fz6c9\") pod \"ovn-controller-ovs-rwzj2\" (UID: \"ea243e32-f3d6-47a6-b2fe-9728af6bc06a\") " pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.529341 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.557924 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:19 crc kubenswrapper[4770]: I1004 04:43:19.715400 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a1a1ae0-84f3-4b49-84e5-1f1737410357" path="/var/lib/kubelet/pods/6a1a1ae0-84f3-4b49-84e5-1f1737410357/volumes" Oct 04 04:43:20 crc kubenswrapper[4770]: I1004 04:43:20.307775 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ncsvk"] Oct 04 04:43:20 crc kubenswrapper[4770]: W1004 04:43:20.513692 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e6ae49d_7530_44a5_bfb1_6201c9e745c8.slice/crio-cb9cc0b96b62b7c8ef6dde371304deeea1fb46b058d3f906acee95f19cb49e2c WatchSource:0}: Error finding container cb9cc0b96b62b7c8ef6dde371304deeea1fb46b058d3f906acee95f19cb49e2c: Status 404 returned error can't find the container with id cb9cc0b96b62b7c8ef6dde371304deeea1fb46b058d3f906acee95f19cb49e2c Oct 04 04:43:20 crc kubenswrapper[4770]: I1004 04:43:20.646156 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rwzj2"] Oct 04 04:43:20 crc kubenswrapper[4770]: W1004 04:43:20.661238 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea243e32_f3d6_47a6_b2fe_9728af6bc06a.slice/crio-e11d06b92bc7b77cc24e079a4bbf27509f305b2c6eafc7901f1b0c66c88418ea WatchSource:0}: Error finding container e11d06b92bc7b77cc24e079a4bbf27509f305b2c6eafc7901f1b0c66c88418ea: Status 404 returned error can't find the container with id e11d06b92bc7b77cc24e079a4bbf27509f305b2c6eafc7901f1b0c66c88418ea Oct 04 04:43:20 crc kubenswrapper[4770]: I1004 04:43:20.933525 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ncsvk" event={"ID":"8e6ae49d-7530-44a5-bfb1-6201c9e745c8","Type":"ContainerStarted","Data":"cb9cc0b96b62b7c8ef6dde371304deeea1fb46b058d3f906acee95f19cb49e2c"} Oct 04 04:43:20 crc kubenswrapper[4770]: I1004 04:43:20.936400 4770 generic.go:334] "Generic (PLEG): container finished" podID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerID="418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee" exitCode=0 Oct 04 04:43:20 crc kubenswrapper[4770]: I1004 04:43:20.936464 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnp42" event={"ID":"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16","Type":"ContainerDied","Data":"418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee"} Oct 04 04:43:20 crc kubenswrapper[4770]: I1004 04:43:20.938252 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rwzj2" event={"ID":"ea243e32-f3d6-47a6-b2fe-9728af6bc06a","Type":"ContainerStarted","Data":"e11d06b92bc7b77cc24e079a4bbf27509f305b2c6eafc7901f1b0c66c88418ea"} Oct 04 04:43:20 crc kubenswrapper[4770]: I1004 04:43:20.945586 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-857xz" event={"ID":"756a79db-b81a-4095-93a3-0c5ecbbaacf6","Type":"ContainerStarted","Data":"d81c5235bc65294e9e1067af096f5f0d59802a75973a42d086055d56dca88470"} Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.821970 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-n4zz4"] Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.823459 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.827208 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.849082 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-n4zz4"] Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.952356 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fcae710-09dc-499b-8bba-fcab1d2b6811-config\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.952428 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5fcae710-09dc-499b-8bba-fcab1d2b6811-ovs-rundir\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.952449 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vdgs\" (UniqueName: \"kubernetes.io/projected/5fcae710-09dc-499b-8bba-fcab1d2b6811-kube-api-access-2vdgs\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.952617 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5fcae710-09dc-499b-8bba-fcab1d2b6811-ovn-rundir\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.956578 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ncsvk" event={"ID":"8e6ae49d-7530-44a5-bfb1-6201c9e745c8","Type":"ContainerStarted","Data":"a2b5669eee0863908763e373d093dfc828946c99253d245f21a7cfe5cb8e67ba"} Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.957669 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.962811 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rwzj2" event={"ID":"ea243e32-f3d6-47a6-b2fe-9728af6bc06a","Type":"ContainerStarted","Data":"ca305043c7fd5000937a6cd8a7b6412055647719a8a08910cbdfc23e7703c9e1"} Oct 04 04:43:21 crc kubenswrapper[4770]: I1004 04:43:21.976903 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ncsvk" podStartSLOduration=2.976887792 podStartE2EDuration="2.976887792s" podCreationTimestamp="2025-10-04 04:43:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:43:21.975327131 +0000 UTC m=+6013.267336843" watchObservedRunningTime="2025-10-04 04:43:21.976887792 +0000 UTC m=+6013.268897504" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.000225 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-857xz" podStartSLOduration=3.694794951 podStartE2EDuration="10.000208972s" podCreationTimestamp="2025-10-04 04:43:12 +0000 UTC" firstStartedPulling="2025-10-04 04:43:13.85115973 +0000 UTC m=+6005.143169442" lastFinishedPulling="2025-10-04 04:43:20.156573751 +0000 UTC m=+6011.448583463" observedRunningTime="2025-10-04 04:43:21.993732603 +0000 UTC m=+6013.285742325" watchObservedRunningTime="2025-10-04 04:43:22.000208972 +0000 UTC m=+6013.292218684" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.053869 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5fcae710-09dc-499b-8bba-fcab1d2b6811-ovn-rundir\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.054027 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fcae710-09dc-499b-8bba-fcab1d2b6811-config\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.054058 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5fcae710-09dc-499b-8bba-fcab1d2b6811-ovs-rundir\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.054077 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vdgs\" (UniqueName: \"kubernetes.io/projected/5fcae710-09dc-499b-8bba-fcab1d2b6811-kube-api-access-2vdgs\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.055283 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5fcae710-09dc-499b-8bba-fcab1d2b6811-ovn-rundir\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.055340 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5fcae710-09dc-499b-8bba-fcab1d2b6811-ovs-rundir\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.056033 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fcae710-09dc-499b-8bba-fcab1d2b6811-config\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.075160 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vdgs\" (UniqueName: \"kubernetes.io/projected/5fcae710-09dc-499b-8bba-fcab1d2b6811-kube-api-access-2vdgs\") pod \"ovn-controller-metrics-n4zz4\" (UID: \"5fcae710-09dc-499b-8bba-fcab1d2b6811\") " pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.160116 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-n4zz4" Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.683697 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-n4zz4"] Oct 04 04:43:22 crc kubenswrapper[4770]: I1004 04:43:22.983560 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-n4zz4" event={"ID":"5fcae710-09dc-499b-8bba-fcab1d2b6811","Type":"ContainerStarted","Data":"b8bfdee700597531645d56cb7fdafacce686534c465c672f3647a7a816f3e04f"} Oct 04 04:43:23 crc kubenswrapper[4770]: I1004 04:43:23.006664 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:23 crc kubenswrapper[4770]: I1004 04:43:23.006929 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:23 crc kubenswrapper[4770]: I1004 04:43:23.316131 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-tr7wv"] Oct 04 04:43:23 crc kubenswrapper[4770]: I1004 04:43:23.317634 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-tr7wv" Oct 04 04:43:23 crc kubenswrapper[4770]: I1004 04:43:23.325160 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-tr7wv"] Oct 04 04:43:23 crc kubenswrapper[4770]: I1004 04:43:23.481218 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgnbd\" (UniqueName: \"kubernetes.io/projected/74462d5e-59ea-4724-bdd5-d68f3640a0ca-kube-api-access-fgnbd\") pod \"octavia-db-create-tr7wv\" (UID: \"74462d5e-59ea-4724-bdd5-d68f3640a0ca\") " pod="openstack/octavia-db-create-tr7wv" Oct 04 04:43:23 crc kubenswrapper[4770]: I1004 04:43:23.583344 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgnbd\" (UniqueName: \"kubernetes.io/projected/74462d5e-59ea-4724-bdd5-d68f3640a0ca-kube-api-access-fgnbd\") pod \"octavia-db-create-tr7wv\" (UID: \"74462d5e-59ea-4724-bdd5-d68f3640a0ca\") " pod="openstack/octavia-db-create-tr7wv" Oct 04 04:43:23 crc kubenswrapper[4770]: I1004 04:43:23.606703 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgnbd\" (UniqueName: \"kubernetes.io/projected/74462d5e-59ea-4724-bdd5-d68f3640a0ca-kube-api-access-fgnbd\") pod \"octavia-db-create-tr7wv\" (UID: \"74462d5e-59ea-4724-bdd5-d68f3640a0ca\") " pod="openstack/octavia-db-create-tr7wv" Oct 04 04:43:23 crc kubenswrapper[4770]: I1004 04:43:23.638260 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-tr7wv" Oct 04 04:43:24 crc kubenswrapper[4770]: I1004 04:43:24.012869 4770 generic.go:334] "Generic (PLEG): container finished" podID="ea243e32-f3d6-47a6-b2fe-9728af6bc06a" containerID="ca305043c7fd5000937a6cd8a7b6412055647719a8a08910cbdfc23e7703c9e1" exitCode=0 Oct 04 04:43:24 crc kubenswrapper[4770]: I1004 04:43:24.013065 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rwzj2" event={"ID":"ea243e32-f3d6-47a6-b2fe-9728af6bc06a","Type":"ContainerDied","Data":"ca305043c7fd5000937a6cd8a7b6412055647719a8a08910cbdfc23e7703c9e1"} Oct 04 04:43:24 crc kubenswrapper[4770]: I1004 04:43:24.017394 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-n4zz4" event={"ID":"5fcae710-09dc-499b-8bba-fcab1d2b6811","Type":"ContainerStarted","Data":"04c32221702daa2d5c192d77e0aad18a4e36b9d46af0acbda2749b3ad115eb86"} Oct 04 04:43:24 crc kubenswrapper[4770]: I1004 04:43:24.056638 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-n4zz4" podStartSLOduration=3.056617588 podStartE2EDuration="3.056617588s" podCreationTimestamp="2025-10-04 04:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:43:24.053351403 +0000 UTC m=+6015.345361115" watchObservedRunningTime="2025-10-04 04:43:24.056617588 +0000 UTC m=+6015.348627310" Oct 04 04:43:24 crc kubenswrapper[4770]: I1004 04:43:24.064262 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-857xz" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="registry-server" probeResult="failure" output=< Oct 04 04:43:24 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 04:43:24 crc kubenswrapper[4770]: > Oct 04 04:43:24 crc kubenswrapper[4770]: I1004 04:43:24.288369 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-tr7wv"] Oct 04 04:43:25 crc kubenswrapper[4770]: I1004 04:43:25.026708 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnp42" event={"ID":"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16","Type":"ContainerStarted","Data":"4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25"} Oct 04 04:43:25 crc kubenswrapper[4770]: I1004 04:43:25.033051 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rwzj2" event={"ID":"ea243e32-f3d6-47a6-b2fe-9728af6bc06a","Type":"ContainerStarted","Data":"84639cd3de4b4e8b01928e8c505a644d94c5cc36e5cbbb3b85eb59f4438c5533"} Oct 04 04:43:25 crc kubenswrapper[4770]: I1004 04:43:25.033124 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rwzj2" event={"ID":"ea243e32-f3d6-47a6-b2fe-9728af6bc06a","Type":"ContainerStarted","Data":"6fe00d8e50f72b12c9ed91726650720dbd5a6e91c18277e27ba75a5beec044be"} Oct 04 04:43:25 crc kubenswrapper[4770]: I1004 04:43:25.033355 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:25 crc kubenswrapper[4770]: I1004 04:43:25.033419 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:25 crc kubenswrapper[4770]: I1004 04:43:25.034729 4770 generic.go:334] "Generic (PLEG): container finished" podID="74462d5e-59ea-4724-bdd5-d68f3640a0ca" containerID="f36650878e20c4c6bc39791f4ad25ad88d047565db4069f8ad0821914e9629cd" exitCode=0 Oct 04 04:43:25 crc kubenswrapper[4770]: I1004 04:43:25.034852 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-tr7wv" event={"ID":"74462d5e-59ea-4724-bdd5-d68f3640a0ca","Type":"ContainerDied","Data":"f36650878e20c4c6bc39791f4ad25ad88d047565db4069f8ad0821914e9629cd"} Oct 04 04:43:25 crc kubenswrapper[4770]: I1004 04:43:25.034899 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-tr7wv" event={"ID":"74462d5e-59ea-4724-bdd5-d68f3640a0ca","Type":"ContainerStarted","Data":"ad95bd3d5eb2754c227878944ed310e3e7ddf2c14465679232f9b44189bcc5c8"} Oct 04 04:43:25 crc kubenswrapper[4770]: I1004 04:43:25.057517 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cnp42" podStartSLOduration=4.174004627 podStartE2EDuration="13.057497446s" podCreationTimestamp="2025-10-04 04:43:12 +0000 UTC" firstStartedPulling="2025-10-04 04:43:14.859950077 +0000 UTC m=+6006.151959809" lastFinishedPulling="2025-10-04 04:43:23.743442916 +0000 UTC m=+6015.035452628" observedRunningTime="2025-10-04 04:43:25.054176299 +0000 UTC m=+6016.346186041" watchObservedRunningTime="2025-10-04 04:43:25.057497446 +0000 UTC m=+6016.349507158" Oct 04 04:43:26 crc kubenswrapper[4770]: I1004 04:43:26.068637 4770 scope.go:117] "RemoveContainer" containerID="8eacfd4e7bd934d5b1e73d8573720bc8711fe1243b893829406c291b13f18666" Oct 04 04:43:26 crc kubenswrapper[4770]: I1004 04:43:26.373844 4770 scope.go:117] "RemoveContainer" containerID="f11cb846bf2694cc3030034642a355c37a0e5ac1410cde3c1a4c2a66ccf0f726" Oct 04 04:43:26 crc kubenswrapper[4770]: I1004 04:43:26.398789 4770 scope.go:117] "RemoveContainer" containerID="77f09aa65fb939fc11c487a147586f640f73308844864de1e7d50f18f101bc74" Oct 04 04:43:26 crc kubenswrapper[4770]: I1004 04:43:26.505618 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-tr7wv" Oct 04 04:43:26 crc kubenswrapper[4770]: I1004 04:43:26.527956 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-rwzj2" podStartSLOduration=7.527939736 podStartE2EDuration="7.527939736s" podCreationTimestamp="2025-10-04 04:43:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:43:25.098382275 +0000 UTC m=+6016.390391987" watchObservedRunningTime="2025-10-04 04:43:26.527939736 +0000 UTC m=+6017.819949448" Oct 04 04:43:26 crc kubenswrapper[4770]: I1004 04:43:26.646506 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgnbd\" (UniqueName: \"kubernetes.io/projected/74462d5e-59ea-4724-bdd5-d68f3640a0ca-kube-api-access-fgnbd\") pod \"74462d5e-59ea-4724-bdd5-d68f3640a0ca\" (UID: \"74462d5e-59ea-4724-bdd5-d68f3640a0ca\") " Oct 04 04:43:26 crc kubenswrapper[4770]: I1004 04:43:26.654534 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74462d5e-59ea-4724-bdd5-d68f3640a0ca-kube-api-access-fgnbd" (OuterVolumeSpecName: "kube-api-access-fgnbd") pod "74462d5e-59ea-4724-bdd5-d68f3640a0ca" (UID: "74462d5e-59ea-4724-bdd5-d68f3640a0ca"). InnerVolumeSpecName "kube-api-access-fgnbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:43:26 crc kubenswrapper[4770]: I1004 04:43:26.748610 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgnbd\" (UniqueName: \"kubernetes.io/projected/74462d5e-59ea-4724-bdd5-d68f3640a0ca-kube-api-access-fgnbd\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:27 crc kubenswrapper[4770]: I1004 04:43:27.057394 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-tr7wv" event={"ID":"74462d5e-59ea-4724-bdd5-d68f3640a0ca","Type":"ContainerDied","Data":"ad95bd3d5eb2754c227878944ed310e3e7ddf2c14465679232f9b44189bcc5c8"} Oct 04 04:43:27 crc kubenswrapper[4770]: I1004 04:43:27.057441 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-tr7wv" Oct 04 04:43:27 crc kubenswrapper[4770]: I1004 04:43:27.057446 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad95bd3d5eb2754c227878944ed310e3e7ddf2c14465679232f9b44189bcc5c8" Oct 04 04:43:31 crc kubenswrapper[4770]: I1004 04:43:31.796515 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:43:31 crc kubenswrapper[4770]: I1004 04:43:31.797324 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:43:33 crc kubenswrapper[4770]: I1004 04:43:33.492348 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:33 crc kubenswrapper[4770]: I1004 04:43:33.492770 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:33 crc kubenswrapper[4770]: I1004 04:43:33.555397 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:34 crc kubenswrapper[4770]: I1004 04:43:34.124110 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-857xz" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="registry-server" probeResult="failure" output=< Oct 04 04:43:34 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 04:43:34 crc kubenswrapper[4770]: > Oct 04 04:43:34 crc kubenswrapper[4770]: I1004 04:43:34.183055 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:34 crc kubenswrapper[4770]: I1004 04:43:34.237245 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cnp42"] Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.062841 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-f8cf-account-create-cst6w"] Oct 04 04:43:36 crc kubenswrapper[4770]: E1004 04:43:36.063650 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74462d5e-59ea-4724-bdd5-d68f3640a0ca" containerName="mariadb-database-create" Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.063663 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="74462d5e-59ea-4724-bdd5-d68f3640a0ca" containerName="mariadb-database-create" Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.063843 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="74462d5e-59ea-4724-bdd5-d68f3640a0ca" containerName="mariadb-database-create" Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.064460 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f8cf-account-create-cst6w" Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.067300 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.086159 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-f8cf-account-create-cst6w"] Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.146418 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k2jr\" (UniqueName: \"kubernetes.io/projected/7d300c24-108b-4b17-a84c-035a0db8b6bc-kube-api-access-7k2jr\") pod \"octavia-f8cf-account-create-cst6w\" (UID: \"7d300c24-108b-4b17-a84c-035a0db8b6bc\") " pod="openstack/octavia-f8cf-account-create-cst6w" Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.165338 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cnp42" podUID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerName="registry-server" containerID="cri-o://4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25" gracePeriod=2 Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.249191 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k2jr\" (UniqueName: \"kubernetes.io/projected/7d300c24-108b-4b17-a84c-035a0db8b6bc-kube-api-access-7k2jr\") pod \"octavia-f8cf-account-create-cst6w\" (UID: \"7d300c24-108b-4b17-a84c-035a0db8b6bc\") " pod="openstack/octavia-f8cf-account-create-cst6w" Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.268343 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k2jr\" (UniqueName: \"kubernetes.io/projected/7d300c24-108b-4b17-a84c-035a0db8b6bc-kube-api-access-7k2jr\") pod \"octavia-f8cf-account-create-cst6w\" (UID: \"7d300c24-108b-4b17-a84c-035a0db8b6bc\") " pod="openstack/octavia-f8cf-account-create-cst6w" Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.387230 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f8cf-account-create-cst6w" Oct 04 04:43:36 crc kubenswrapper[4770]: I1004 04:43:36.844394 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-f8cf-account-create-cst6w"] Oct 04 04:43:36 crc kubenswrapper[4770]: W1004 04:43:36.847464 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d300c24_108b_4b17_a84c_035a0db8b6bc.slice/crio-c5c08420e31f254af74319c2fdbc894a5093f182bb5eb3f947394d0e816c4be7 WatchSource:0}: Error finding container c5c08420e31f254af74319c2fdbc894a5093f182bb5eb3f947394d0e816c4be7: Status 404 returned error can't find the container with id c5c08420e31f254af74319c2fdbc894a5093f182bb5eb3f947394d0e816c4be7 Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.149061 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.183988 4770 generic.go:334] "Generic (PLEG): container finished" podID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerID="4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25" exitCode=0 Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.184075 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnp42" event={"ID":"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16","Type":"ContainerDied","Data":"4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25"} Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.184103 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnp42" event={"ID":"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16","Type":"ContainerDied","Data":"7c6422c2f761dcb29852714227d067ff81fd214091f50bf64e509c9124a5d427"} Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.184119 4770 scope.go:117] "RemoveContainer" containerID="4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.184228 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnp42" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.187492 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f8cf-account-create-cst6w" event={"ID":"7d300c24-108b-4b17-a84c-035a0db8b6bc","Type":"ContainerStarted","Data":"cce914e63f656abc95c2d67af215f914022cec3d6c94b3b1ca8ee891fd5fb80f"} Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.187528 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f8cf-account-create-cst6w" event={"ID":"7d300c24-108b-4b17-a84c-035a0db8b6bc","Type":"ContainerStarted","Data":"c5c08420e31f254af74319c2fdbc894a5093f182bb5eb3f947394d0e816c4be7"} Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.208652 4770 scope.go:117] "RemoveContainer" containerID="418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.244122 4770 scope.go:117] "RemoveContainer" containerID="d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.263516 4770 scope.go:117] "RemoveContainer" containerID="4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25" Oct 04 04:43:37 crc kubenswrapper[4770]: E1004 04:43:37.264029 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25\": container with ID starting with 4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25 not found: ID does not exist" containerID="4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.264066 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25"} err="failed to get container status \"4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25\": rpc error: code = NotFound desc = could not find container \"4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25\": container with ID starting with 4e4eb140fc112b6a58860f4b80125167bd2ad7ec1f5004bf1183bcea8f026b25 not found: ID does not exist" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.264095 4770 scope.go:117] "RemoveContainer" containerID="418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee" Oct 04 04:43:37 crc kubenswrapper[4770]: E1004 04:43:37.264405 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee\": container with ID starting with 418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee not found: ID does not exist" containerID="418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.264441 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee"} err="failed to get container status \"418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee\": rpc error: code = NotFound desc = could not find container \"418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee\": container with ID starting with 418c08204dbd15e259d92b9525565f99a65fcb14d5c56f323926e6df452aa9ee not found: ID does not exist" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.264457 4770 scope.go:117] "RemoveContainer" containerID="d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615" Oct 04 04:43:37 crc kubenswrapper[4770]: E1004 04:43:37.265354 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615\": container with ID starting with d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615 not found: ID does not exist" containerID="d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.265386 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615"} err="failed to get container status \"d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615\": rpc error: code = NotFound desc = could not find container \"d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615\": container with ID starting with d47912583e20773798911fbf37bd7685556169270b54b55c08a3386653546615 not found: ID does not exist" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.273246 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzrc6\" (UniqueName: \"kubernetes.io/projected/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-kube-api-access-dzrc6\") pod \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.273415 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-utilities\") pod \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.273450 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-catalog-content\") pod \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\" (UID: \"0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16\") " Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.276527 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-utilities" (OuterVolumeSpecName: "utilities") pod "0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" (UID: "0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.281739 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-kube-api-access-dzrc6" (OuterVolumeSpecName: "kube-api-access-dzrc6") pod "0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" (UID: "0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16"). InnerVolumeSpecName "kube-api-access-dzrc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.328487 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" (UID: "0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.375627 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.375661 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzrc6\" (UniqueName: \"kubernetes.io/projected/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-kube-api-access-dzrc6\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.375676 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.552281 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cnp42"] Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.561167 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cnp42"] Oct 04 04:43:37 crc kubenswrapper[4770]: I1004 04:43:37.697239 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" path="/var/lib/kubelet/pods/0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16/volumes" Oct 04 04:43:38 crc kubenswrapper[4770]: I1004 04:43:38.203591 4770 generic.go:334] "Generic (PLEG): container finished" podID="7d300c24-108b-4b17-a84c-035a0db8b6bc" containerID="cce914e63f656abc95c2d67af215f914022cec3d6c94b3b1ca8ee891fd5fb80f" exitCode=0 Oct 04 04:43:38 crc kubenswrapper[4770]: I1004 04:43:38.203636 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f8cf-account-create-cst6w" event={"ID":"7d300c24-108b-4b17-a84c-035a0db8b6bc","Type":"ContainerDied","Data":"cce914e63f656abc95c2d67af215f914022cec3d6c94b3b1ca8ee891fd5fb80f"} Oct 04 04:43:39 crc kubenswrapper[4770]: I1004 04:43:39.577986 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f8cf-account-create-cst6w" Oct 04 04:43:39 crc kubenswrapper[4770]: I1004 04:43:39.728822 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k2jr\" (UniqueName: \"kubernetes.io/projected/7d300c24-108b-4b17-a84c-035a0db8b6bc-kube-api-access-7k2jr\") pod \"7d300c24-108b-4b17-a84c-035a0db8b6bc\" (UID: \"7d300c24-108b-4b17-a84c-035a0db8b6bc\") " Oct 04 04:43:39 crc kubenswrapper[4770]: I1004 04:43:39.736337 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d300c24-108b-4b17-a84c-035a0db8b6bc-kube-api-access-7k2jr" (OuterVolumeSpecName: "kube-api-access-7k2jr") pod "7d300c24-108b-4b17-a84c-035a0db8b6bc" (UID: "7d300c24-108b-4b17-a84c-035a0db8b6bc"). InnerVolumeSpecName "kube-api-access-7k2jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:43:39 crc kubenswrapper[4770]: I1004 04:43:39.831527 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k2jr\" (UniqueName: \"kubernetes.io/projected/7d300c24-108b-4b17-a84c-035a0db8b6bc-kube-api-access-7k2jr\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:40 crc kubenswrapper[4770]: I1004 04:43:40.233677 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-f8cf-account-create-cst6w" event={"ID":"7d300c24-108b-4b17-a84c-035a0db8b6bc","Type":"ContainerDied","Data":"c5c08420e31f254af74319c2fdbc894a5093f182bb5eb3f947394d0e816c4be7"} Oct 04 04:43:40 crc kubenswrapper[4770]: I1004 04:43:40.233953 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5c08420e31f254af74319c2fdbc894a5093f182bb5eb3f947394d0e816c4be7" Oct 04 04:43:40 crc kubenswrapper[4770]: I1004 04:43:40.233768 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-f8cf-account-create-cst6w" Oct 04 04:43:41 crc kubenswrapper[4770]: I1004 04:43:41.889348 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-5kgpr"] Oct 04 04:43:41 crc kubenswrapper[4770]: E1004 04:43:41.890267 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerName="extract-content" Oct 04 04:43:41 crc kubenswrapper[4770]: I1004 04:43:41.890284 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerName="extract-content" Oct 04 04:43:41 crc kubenswrapper[4770]: E1004 04:43:41.890322 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d300c24-108b-4b17-a84c-035a0db8b6bc" containerName="mariadb-account-create" Oct 04 04:43:41 crc kubenswrapper[4770]: I1004 04:43:41.890328 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d300c24-108b-4b17-a84c-035a0db8b6bc" containerName="mariadb-account-create" Oct 04 04:43:41 crc kubenswrapper[4770]: E1004 04:43:41.890341 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerName="extract-utilities" Oct 04 04:43:41 crc kubenswrapper[4770]: I1004 04:43:41.890348 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerName="extract-utilities" Oct 04 04:43:41 crc kubenswrapper[4770]: E1004 04:43:41.890367 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerName="registry-server" Oct 04 04:43:41 crc kubenswrapper[4770]: I1004 04:43:41.890373 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerName="registry-server" Oct 04 04:43:41 crc kubenswrapper[4770]: I1004 04:43:41.890581 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dab3d1c-4281-4d29-b9b2-f86e4b0e5a16" containerName="registry-server" Oct 04 04:43:41 crc kubenswrapper[4770]: I1004 04:43:41.890604 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d300c24-108b-4b17-a84c-035a0db8b6bc" containerName="mariadb-account-create" Oct 04 04:43:41 crc kubenswrapper[4770]: I1004 04:43:41.891326 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-5kgpr" Oct 04 04:43:41 crc kubenswrapper[4770]: I1004 04:43:41.902553 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-5kgpr"] Oct 04 04:43:42 crc kubenswrapper[4770]: I1004 04:43:42.072571 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ghhf\" (UniqueName: \"kubernetes.io/projected/bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1-kube-api-access-5ghhf\") pod \"octavia-persistence-db-create-5kgpr\" (UID: \"bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1\") " pod="openstack/octavia-persistence-db-create-5kgpr" Oct 04 04:43:42 crc kubenswrapper[4770]: I1004 04:43:42.174556 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ghhf\" (UniqueName: \"kubernetes.io/projected/bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1-kube-api-access-5ghhf\") pod \"octavia-persistence-db-create-5kgpr\" (UID: \"bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1\") " pod="openstack/octavia-persistence-db-create-5kgpr" Oct 04 04:43:42 crc kubenswrapper[4770]: I1004 04:43:42.202187 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ghhf\" (UniqueName: \"kubernetes.io/projected/bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1-kube-api-access-5ghhf\") pod \"octavia-persistence-db-create-5kgpr\" (UID: \"bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1\") " pod="openstack/octavia-persistence-db-create-5kgpr" Oct 04 04:43:42 crc kubenswrapper[4770]: I1004 04:43:42.220497 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-5kgpr" Oct 04 04:43:42 crc kubenswrapper[4770]: I1004 04:43:42.737976 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-5kgpr"] Oct 04 04:43:43 crc kubenswrapper[4770]: I1004 04:43:43.271947 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-5kgpr" event={"ID":"bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1","Type":"ContainerStarted","Data":"d6f4a444c09ba05fa6e129d32ef43d4320ee28d7ba3b0b8a19592d9da8c7accb"} Oct 04 04:43:43 crc kubenswrapper[4770]: I1004 04:43:43.272251 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-5kgpr" event={"ID":"bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1","Type":"ContainerStarted","Data":"8b5b11a93ddb55cd1c9d600f32dea674858e01eb64259ed175d890d63fe3b804"} Oct 04 04:43:44 crc kubenswrapper[4770]: I1004 04:43:44.052435 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-857xz" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="registry-server" probeResult="failure" output=< Oct 04 04:43:44 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 04:43:44 crc kubenswrapper[4770]: > Oct 04 04:43:44 crc kubenswrapper[4770]: I1004 04:43:44.285115 4770 generic.go:334] "Generic (PLEG): container finished" podID="bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1" containerID="d6f4a444c09ba05fa6e129d32ef43d4320ee28d7ba3b0b8a19592d9da8c7accb" exitCode=0 Oct 04 04:43:44 crc kubenswrapper[4770]: I1004 04:43:44.285401 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-5kgpr" event={"ID":"bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1","Type":"ContainerDied","Data":"d6f4a444c09ba05fa6e129d32ef43d4320ee28d7ba3b0b8a19592d9da8c7accb"} Oct 04 04:43:45 crc kubenswrapper[4770]: I1004 04:43:45.643953 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-5kgpr" Oct 04 04:43:45 crc kubenswrapper[4770]: I1004 04:43:45.665616 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ghhf\" (UniqueName: \"kubernetes.io/projected/bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1-kube-api-access-5ghhf\") pod \"bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1\" (UID: \"bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1\") " Oct 04 04:43:45 crc kubenswrapper[4770]: I1004 04:43:45.671847 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1-kube-api-access-5ghhf" (OuterVolumeSpecName: "kube-api-access-5ghhf") pod "bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1" (UID: "bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1"). InnerVolumeSpecName "kube-api-access-5ghhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:43:45 crc kubenswrapper[4770]: I1004 04:43:45.767992 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ghhf\" (UniqueName: \"kubernetes.io/projected/bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1-kube-api-access-5ghhf\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:46 crc kubenswrapper[4770]: I1004 04:43:46.304121 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-5kgpr" event={"ID":"bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1","Type":"ContainerDied","Data":"8b5b11a93ddb55cd1c9d600f32dea674858e01eb64259ed175d890d63fe3b804"} Oct 04 04:43:46 crc kubenswrapper[4770]: I1004 04:43:46.304495 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b5b11a93ddb55cd1c9d600f32dea674858e01eb64259ed175d890d63fe3b804" Oct 04 04:43:46 crc kubenswrapper[4770]: I1004 04:43:46.304198 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-5kgpr" Oct 04 04:43:49 crc kubenswrapper[4770]: I1004 04:43:49.055949 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-d2nkz"] Oct 04 04:43:49 crc kubenswrapper[4770]: I1004 04:43:49.064881 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-d2nkz"] Oct 04 04:43:49 crc kubenswrapper[4770]: I1004 04:43:49.689990 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70d9f4e9-231d-49fe-b7ac-ac9f9301e501" path="/var/lib/kubelet/pods/70d9f4e9-231d-49fe-b7ac-ac9f9301e501/volumes" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.083318 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.149565 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.325636 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-857xz"] Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.414193 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-da4f-account-create-l27vj"] Oct 04 04:43:53 crc kubenswrapper[4770]: E1004 04:43:53.414745 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1" containerName="mariadb-database-create" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.414769 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1" containerName="mariadb-database-create" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.415054 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1" containerName="mariadb-database-create" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.415760 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-da4f-account-create-l27vj" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.418176 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.422943 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-da4f-account-create-l27vj"] Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.552949 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc62c\" (UniqueName: \"kubernetes.io/projected/dd207fcd-44df-4b1b-bff9-630f556d7a5f-kube-api-access-gc62c\") pod \"octavia-da4f-account-create-l27vj\" (UID: \"dd207fcd-44df-4b1b-bff9-630f556d7a5f\") " pod="openstack/octavia-da4f-account-create-l27vj" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.654794 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc62c\" (UniqueName: \"kubernetes.io/projected/dd207fcd-44df-4b1b-bff9-630f556d7a5f-kube-api-access-gc62c\") pod \"octavia-da4f-account-create-l27vj\" (UID: \"dd207fcd-44df-4b1b-bff9-630f556d7a5f\") " pod="openstack/octavia-da4f-account-create-l27vj" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.686499 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc62c\" (UniqueName: \"kubernetes.io/projected/dd207fcd-44df-4b1b-bff9-630f556d7a5f-kube-api-access-gc62c\") pod \"octavia-da4f-account-create-l27vj\" (UID: \"dd207fcd-44df-4b1b-bff9-630f556d7a5f\") " pod="openstack/octavia-da4f-account-create-l27vj" Oct 04 04:43:53 crc kubenswrapper[4770]: I1004 04:43:53.739816 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-da4f-account-create-l27vj" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.233030 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-da4f-account-create-l27vj"] Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.399769 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-857xz" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="registry-server" containerID="cri-o://d81c5235bc65294e9e1067af096f5f0d59802a75973a42d086055d56dca88470" gracePeriod=2 Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.400507 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-da4f-account-create-l27vj" event={"ID":"dd207fcd-44df-4b1b-bff9-630f556d7a5f","Type":"ContainerStarted","Data":"6784b240d584fb1fff881b0f0d729800918508a49bc986b213a8a47ae951a7cd"} Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.578847 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ncsvk" podUID="8e6ae49d-7530-44a5-bfb1-6201c9e745c8" containerName="ovn-controller" probeResult="failure" output=< Oct 04 04:43:54 crc kubenswrapper[4770]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 04:43:54 crc kubenswrapper[4770]: > Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.626318 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.637797 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rwzj2" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.783335 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ncsvk-config-cwdnx"] Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.785218 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.787697 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.792349 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ncsvk-config-cwdnx"] Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.981098 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.981168 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run-ovn\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.981370 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-scripts\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.981433 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-additional-scripts\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.981467 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v28rv\" (UniqueName: \"kubernetes.io/projected/211abe0e-8a56-4f08-ad2a-8b1a174e6433-kube-api-access-v28rv\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:54 crc kubenswrapper[4770]: I1004 04:43:54.981538 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-log-ovn\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.083574 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.083638 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run-ovn\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.083689 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-scripts\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.083711 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-additional-scripts\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.083729 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v28rv\" (UniqueName: \"kubernetes.io/projected/211abe0e-8a56-4f08-ad2a-8b1a174e6433-kube-api-access-v28rv\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.083755 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-log-ovn\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.083962 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.084020 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-log-ovn\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.084028 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run-ovn\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.084636 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-additional-scripts\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.087074 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-scripts\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.108617 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v28rv\" (UniqueName: \"kubernetes.io/projected/211abe0e-8a56-4f08-ad2a-8b1a174e6433-kube-api-access-v28rv\") pod \"ovn-controller-ncsvk-config-cwdnx\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.405741 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.410236 4770 generic.go:334] "Generic (PLEG): container finished" podID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerID="d81c5235bc65294e9e1067af096f5f0d59802a75973a42d086055d56dca88470" exitCode=0 Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.410301 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-857xz" event={"ID":"756a79db-b81a-4095-93a3-0c5ecbbaacf6","Type":"ContainerDied","Data":"d81c5235bc65294e9e1067af096f5f0d59802a75973a42d086055d56dca88470"} Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.411450 4770 generic.go:334] "Generic (PLEG): container finished" podID="dd207fcd-44df-4b1b-bff9-630f556d7a5f" containerID="ff2b58e303ddc23248e6155a5eae386811b9a87ada716158f0026065b9474e5a" exitCode=0 Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.412836 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-da4f-account-create-l27vj" event={"ID":"dd207fcd-44df-4b1b-bff9-630f556d7a5f","Type":"ContainerDied","Data":"ff2b58e303ddc23248e6155a5eae386811b9a87ada716158f0026065b9474e5a"} Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.857878 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:55 crc kubenswrapper[4770]: I1004 04:43:55.861801 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ncsvk-config-cwdnx"] Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.005676 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-utilities\") pod \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.006163 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-catalog-content\") pod \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.006299 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj7xv\" (UniqueName: \"kubernetes.io/projected/756a79db-b81a-4095-93a3-0c5ecbbaacf6-kube-api-access-dj7xv\") pod \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\" (UID: \"756a79db-b81a-4095-93a3-0c5ecbbaacf6\") " Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.006936 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-utilities" (OuterVolumeSpecName: "utilities") pod "756a79db-b81a-4095-93a3-0c5ecbbaacf6" (UID: "756a79db-b81a-4095-93a3-0c5ecbbaacf6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.015273 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/756a79db-b81a-4095-93a3-0c5ecbbaacf6-kube-api-access-dj7xv" (OuterVolumeSpecName: "kube-api-access-dj7xv") pod "756a79db-b81a-4095-93a3-0c5ecbbaacf6" (UID: "756a79db-b81a-4095-93a3-0c5ecbbaacf6"). InnerVolumeSpecName "kube-api-access-dj7xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.052401 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "756a79db-b81a-4095-93a3-0c5ecbbaacf6" (UID: "756a79db-b81a-4095-93a3-0c5ecbbaacf6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.108951 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.108985 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/756a79db-b81a-4095-93a3-0c5ecbbaacf6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.108997 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj7xv\" (UniqueName: \"kubernetes.io/projected/756a79db-b81a-4095-93a3-0c5ecbbaacf6-kube-api-access-dj7xv\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.422989 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-857xz" event={"ID":"756a79db-b81a-4095-93a3-0c5ecbbaacf6","Type":"ContainerDied","Data":"6dae64bb9b321e25c502effac9cca7614beae1e19dc6c430d84a5b187f020769"} Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.423065 4770 scope.go:117] "RemoveContainer" containerID="d81c5235bc65294e9e1067af096f5f0d59802a75973a42d086055d56dca88470" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.423131 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-857xz" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.426828 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ncsvk-config-cwdnx" event={"ID":"211abe0e-8a56-4f08-ad2a-8b1a174e6433","Type":"ContainerStarted","Data":"cf45077d3fae92725c35948ec5d905fd1cc36c0201520ecfc5e506505dc88a89"} Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.426897 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ncsvk-config-cwdnx" event={"ID":"211abe0e-8a56-4f08-ad2a-8b1a174e6433","Type":"ContainerStarted","Data":"6377b3346b51ad39fcd7e1babc2c78c52103621d87c41e2b71b252611dc3e453"} Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.453768 4770 scope.go:117] "RemoveContainer" containerID="d6fe3268d09a0b69fc11d30ef57314bfa20fba09100e52d7002f2a00fc56c449" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.477662 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-857xz"] Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.489366 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-857xz"] Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.492872 4770 scope.go:117] "RemoveContainer" containerID="86254b538369fdc5053b8cb9ca177a51d4117311c37a7c3d3ced2beafe0b0058" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.724959 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-da4f-account-create-l27vj" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.821146 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc62c\" (UniqueName: \"kubernetes.io/projected/dd207fcd-44df-4b1b-bff9-630f556d7a5f-kube-api-access-gc62c\") pod \"dd207fcd-44df-4b1b-bff9-630f556d7a5f\" (UID: \"dd207fcd-44df-4b1b-bff9-630f556d7a5f\") " Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.827885 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd207fcd-44df-4b1b-bff9-630f556d7a5f-kube-api-access-gc62c" (OuterVolumeSpecName: "kube-api-access-gc62c") pod "dd207fcd-44df-4b1b-bff9-630f556d7a5f" (UID: "dd207fcd-44df-4b1b-bff9-630f556d7a5f"). InnerVolumeSpecName "kube-api-access-gc62c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:43:56 crc kubenswrapper[4770]: I1004 04:43:56.923397 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc62c\" (UniqueName: \"kubernetes.io/projected/dd207fcd-44df-4b1b-bff9-630f556d7a5f-kube-api-access-gc62c\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:57 crc kubenswrapper[4770]: I1004 04:43:57.443382 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-da4f-account-create-l27vj" event={"ID":"dd207fcd-44df-4b1b-bff9-630f556d7a5f","Type":"ContainerDied","Data":"6784b240d584fb1fff881b0f0d729800918508a49bc986b213a8a47ae951a7cd"} Oct 04 04:43:57 crc kubenswrapper[4770]: I1004 04:43:57.443692 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6784b240d584fb1fff881b0f0d729800918508a49bc986b213a8a47ae951a7cd" Oct 04 04:43:57 crc kubenswrapper[4770]: I1004 04:43:57.443414 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-da4f-account-create-l27vj" Oct 04 04:43:57 crc kubenswrapper[4770]: I1004 04:43:57.445458 4770 generic.go:334] "Generic (PLEG): container finished" podID="211abe0e-8a56-4f08-ad2a-8b1a174e6433" containerID="cf45077d3fae92725c35948ec5d905fd1cc36c0201520ecfc5e506505dc88a89" exitCode=0 Oct 04 04:43:57 crc kubenswrapper[4770]: I1004 04:43:57.445501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ncsvk-config-cwdnx" event={"ID":"211abe0e-8a56-4f08-ad2a-8b1a174e6433","Type":"ContainerDied","Data":"cf45077d3fae92725c35948ec5d905fd1cc36c0201520ecfc5e506505dc88a89"} Oct 04 04:43:57 crc kubenswrapper[4770]: I1004 04:43:57.693626 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" path="/var/lib/kubelet/pods/756a79db-b81a-4095-93a3-0c5ecbbaacf6/volumes" Oct 04 04:43:58 crc kubenswrapper[4770]: I1004 04:43:58.828377 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.016454 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-scripts\") pod \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.016517 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-log-ovn\") pod \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.016542 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-additional-scripts\") pod \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.016586 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run-ovn\") pod \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.016638 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v28rv\" (UniqueName: \"kubernetes.io/projected/211abe0e-8a56-4f08-ad2a-8b1a174e6433-kube-api-access-v28rv\") pod \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.016676 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run\") pod \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\" (UID: \"211abe0e-8a56-4f08-ad2a-8b1a174e6433\") " Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.017001 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run" (OuterVolumeSpecName: "var-run") pod "211abe0e-8a56-4f08-ad2a-8b1a174e6433" (UID: "211abe0e-8a56-4f08-ad2a-8b1a174e6433"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.017382 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "211abe0e-8a56-4f08-ad2a-8b1a174e6433" (UID: "211abe0e-8a56-4f08-ad2a-8b1a174e6433"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.018037 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "211abe0e-8a56-4f08-ad2a-8b1a174e6433" (UID: "211abe0e-8a56-4f08-ad2a-8b1a174e6433"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.018082 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "211abe0e-8a56-4f08-ad2a-8b1a174e6433" (UID: "211abe0e-8a56-4f08-ad2a-8b1a174e6433"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.018058 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-scripts" (OuterVolumeSpecName: "scripts") pod "211abe0e-8a56-4f08-ad2a-8b1a174e6433" (UID: "211abe0e-8a56-4f08-ad2a-8b1a174e6433"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.024633 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/211abe0e-8a56-4f08-ad2a-8b1a174e6433-kube-api-access-v28rv" (OuterVolumeSpecName: "kube-api-access-v28rv") pod "211abe0e-8a56-4f08-ad2a-8b1a174e6433" (UID: "211abe0e-8a56-4f08-ad2a-8b1a174e6433"). InnerVolumeSpecName "kube-api-access-v28rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.039000 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f532-account-create-q7svr"] Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.046314 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f532-account-create-q7svr"] Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.119939 4770 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.119992 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v28rv\" (UniqueName: \"kubernetes.io/projected/211abe0e-8a56-4f08-ad2a-8b1a174e6433-kube-api-access-v28rv\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.120065 4770 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.120093 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.120115 4770 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/211abe0e-8a56-4f08-ad2a-8b1a174e6433-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.120140 4770 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/211abe0e-8a56-4f08-ad2a-8b1a174e6433-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.484585 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ncsvk-config-cwdnx" event={"ID":"211abe0e-8a56-4f08-ad2a-8b1a174e6433","Type":"ContainerDied","Data":"6377b3346b51ad39fcd7e1babc2c78c52103621d87c41e2b71b252611dc3e453"} Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.484669 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6377b3346b51ad39fcd7e1babc2c78c52103621d87c41e2b71b252611dc3e453" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.484760 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ncsvk-config-cwdnx" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.599905 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ncsvk" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.691646 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0a8eb2-d199-44ed-8dba-0b654c4261ca" path="/var/lib/kubelet/pods/fc0a8eb2-d199-44ed-8dba-0b654c4261ca/volumes" Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.900289 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ncsvk-config-cwdnx"] Oct 04 04:43:59 crc kubenswrapper[4770]: I1004 04:43:59.917837 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ncsvk-config-cwdnx"] Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.041880 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-5968f659f7-n2kjd"] Oct 04 04:44:00 crc kubenswrapper[4770]: E1004 04:44:00.042597 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd207fcd-44df-4b1b-bff9-630f556d7a5f" containerName="mariadb-account-create" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.042615 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd207fcd-44df-4b1b-bff9-630f556d7a5f" containerName="mariadb-account-create" Oct 04 04:44:00 crc kubenswrapper[4770]: E1004 04:44:00.042636 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="211abe0e-8a56-4f08-ad2a-8b1a174e6433" containerName="ovn-config" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.042644 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="211abe0e-8a56-4f08-ad2a-8b1a174e6433" containerName="ovn-config" Oct 04 04:44:00 crc kubenswrapper[4770]: E1004 04:44:00.042672 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="extract-content" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.042681 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="extract-content" Oct 04 04:44:00 crc kubenswrapper[4770]: E1004 04:44:00.042694 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="extract-utilities" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.042701 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="extract-utilities" Oct 04 04:44:00 crc kubenswrapper[4770]: E1004 04:44:00.042714 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="registry-server" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.042722 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="registry-server" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.042977 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="756a79db-b81a-4095-93a3-0c5ecbbaacf6" containerName="registry-server" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.042991 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd207fcd-44df-4b1b-bff9-630f556d7a5f" containerName="mariadb-account-create" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.043000 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="211abe0e-8a56-4f08-ad2a-8b1a174e6433" containerName="ovn-config" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.044820 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.046819 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-7hw2q" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.046976 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.048211 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.059484 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-5968f659f7-n2kjd"] Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.147553 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/198ec713-5ae0-4e0a-ba62-433810e9e486-config-data-merged\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.147677 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198ec713-5ae0-4e0a-ba62-433810e9e486-config-data\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.147815 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198ec713-5ae0-4e0a-ba62-433810e9e486-combined-ca-bundle\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.147886 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/198ec713-5ae0-4e0a-ba62-433810e9e486-octavia-run\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.147930 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198ec713-5ae0-4e0a-ba62-433810e9e486-scripts\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.250385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198ec713-5ae0-4e0a-ba62-433810e9e486-combined-ca-bundle\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.250464 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/198ec713-5ae0-4e0a-ba62-433810e9e486-octavia-run\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.250490 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198ec713-5ae0-4e0a-ba62-433810e9e486-scripts\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.250545 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/198ec713-5ae0-4e0a-ba62-433810e9e486-config-data-merged\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.250588 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198ec713-5ae0-4e0a-ba62-433810e9e486-config-data\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.251047 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/198ec713-5ae0-4e0a-ba62-433810e9e486-octavia-run\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.251335 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/198ec713-5ae0-4e0a-ba62-433810e9e486-config-data-merged\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.258821 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198ec713-5ae0-4e0a-ba62-433810e9e486-combined-ca-bundle\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.258952 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198ec713-5ae0-4e0a-ba62-433810e9e486-scripts\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.260032 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198ec713-5ae0-4e0a-ba62-433810e9e486-config-data\") pod \"octavia-api-5968f659f7-n2kjd\" (UID: \"198ec713-5ae0-4e0a-ba62-433810e9e486\") " pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.362623 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:00 crc kubenswrapper[4770]: I1004 04:44:00.837410 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-5968f659f7-n2kjd"] Oct 04 04:44:00 crc kubenswrapper[4770]: W1004 04:44:00.837779 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod198ec713_5ae0_4e0a_ba62_433810e9e486.slice/crio-ec8743fe72fbb601d952184e051cbf42c45f75daf4eb91a62dcdbd9224050ea6 WatchSource:0}: Error finding container ec8743fe72fbb601d952184e051cbf42c45f75daf4eb91a62dcdbd9224050ea6: Status 404 returned error can't find the container with id ec8743fe72fbb601d952184e051cbf42c45f75daf4eb91a62dcdbd9224050ea6 Oct 04 04:44:01 crc kubenswrapper[4770]: I1004 04:44:01.515497 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5968f659f7-n2kjd" event={"ID":"198ec713-5ae0-4e0a-ba62-433810e9e486","Type":"ContainerStarted","Data":"ec8743fe72fbb601d952184e051cbf42c45f75daf4eb91a62dcdbd9224050ea6"} Oct 04 04:44:01 crc kubenswrapper[4770]: I1004 04:44:01.690383 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="211abe0e-8a56-4f08-ad2a-8b1a174e6433" path="/var/lib/kubelet/pods/211abe0e-8a56-4f08-ad2a-8b1a174e6433/volumes" Oct 04 04:44:01 crc kubenswrapper[4770]: I1004 04:44:01.795974 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:44:01 crc kubenswrapper[4770]: I1004 04:44:01.796076 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:44:11 crc kubenswrapper[4770]: I1004 04:44:11.039194 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-dvh5k"] Oct 04 04:44:11 crc kubenswrapper[4770]: I1004 04:44:11.052178 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-dvh5k"] Oct 04 04:44:11 crc kubenswrapper[4770]: I1004 04:44:11.687415 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33d878f0-d662-4c1c-8fc4-49ccf2f7afb4" path="/var/lib/kubelet/pods/33d878f0-d662-4c1c-8fc4-49ccf2f7afb4/volumes" Oct 04 04:44:20 crc kubenswrapper[4770]: E1004 04:44:20.737334 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:6e7fc684bf020d3552774ed9a2033634e54e0c946741500f3b93c83a0ede01cd" Oct 04 04:44:20 crc kubenswrapper[4770]: E1004 04:44:20.738149 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:6e7fc684bf020d3552774ed9a2033634e54e0c946741500f3b93c83a0ede01cd,Command:[/bin/bash],Args:[-c /usr/local/bin/container-scripts/init.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-merged,ReadOnly:false,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42437,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42437,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-api-5968f659f7-n2kjd_openstack(198ec713-5ae0-4e0a-ba62-433810e9e486): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:44:20 crc kubenswrapper[4770]: E1004 04:44:20.739346 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/octavia-api-5968f659f7-n2kjd" podUID="198ec713-5ae0-4e0a-ba62-433810e9e486" Oct 04 04:44:20 crc kubenswrapper[4770]: E1004 04:44:20.761843 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:6e7fc684bf020d3552774ed9a2033634e54e0c946741500f3b93c83a0ede01cd\\\"\"" pod="openstack/octavia-api-5968f659f7-n2kjd" podUID="198ec713-5ae0-4e0a-ba62-433810e9e486" Oct 04 04:44:26 crc kubenswrapper[4770]: I1004 04:44:26.631266 4770 scope.go:117] "RemoveContainer" containerID="5acac5e1b13ae75d025efdeabbc5171e0281f404b3b0edfb3f5c0fbdab71671a" Oct 04 04:44:26 crc kubenswrapper[4770]: I1004 04:44:26.766069 4770 scope.go:117] "RemoveContainer" containerID="ce09c6ff968307577b4a88a60a9bbc1c71a073fa3b1ab7881ef48a87abff3862" Oct 04 04:44:26 crc kubenswrapper[4770]: I1004 04:44:26.913312 4770 scope.go:117] "RemoveContainer" containerID="76bc11893d6003cdc5a3567ab4d0e9505654d15686caad6c1128642d03b94d4f" Oct 04 04:44:31 crc kubenswrapper[4770]: I1004 04:44:31.795442 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:44:31 crc kubenswrapper[4770]: I1004 04:44:31.796331 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:44:31 crc kubenswrapper[4770]: I1004 04:44:31.796407 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:44:31 crc kubenswrapper[4770]: I1004 04:44:31.797925 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:44:31 crc kubenswrapper[4770]: I1004 04:44:31.798053 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" gracePeriod=600 Oct 04 04:44:31 crc kubenswrapper[4770]: E1004 04:44:31.993207 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bd3be93_1791_4cd1_b3ae_b4032548e93a.slice/crio-272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:44:33 crc kubenswrapper[4770]: I1004 04:44:33.921592 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" exitCode=0 Oct 04 04:44:33 crc kubenswrapper[4770]: I1004 04:44:33.921679 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545"} Oct 04 04:44:33 crc kubenswrapper[4770]: I1004 04:44:33.922035 4770 scope.go:117] "RemoveContainer" containerID="2e30f9d379ccb3500c52f46d443a65b8cc328fea98419caf04a4e9a30d4bd0cb" Oct 04 04:44:35 crc kubenswrapper[4770]: E1004 04:44:35.470815 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:44:35 crc kubenswrapper[4770]: I1004 04:44:35.945248 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:44:35 crc kubenswrapper[4770]: E1004 04:44:35.946458 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:44:38 crc kubenswrapper[4770]: I1004 04:44:38.826538 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-mjbrc"] Oct 04 04:44:38 crc kubenswrapper[4770]: I1004 04:44:38.830793 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:38 crc kubenswrapper[4770]: I1004 04:44:38.833135 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Oct 04 04:44:38 crc kubenswrapper[4770]: I1004 04:44:38.836694 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 04 04:44:38 crc kubenswrapper[4770]: I1004 04:44:38.850393 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-mjbrc"] Oct 04 04:44:38 crc kubenswrapper[4770]: I1004 04:44:38.923918 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-scripts\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:38 crc kubenswrapper[4770]: I1004 04:44:38.923991 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/109f402b-830b-4988-9a1f-f1d56652be2c-config-data-merged\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:38 crc kubenswrapper[4770]: I1004 04:44:38.924043 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-combined-ca-bundle\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:38 crc kubenswrapper[4770]: I1004 04:44:38.924247 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-config-data\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.026185 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-config-data\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.026314 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-scripts\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.026359 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/109f402b-830b-4988-9a1f-f1d56652be2c-config-data-merged\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.026385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-combined-ca-bundle\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.027210 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/109f402b-830b-4988-9a1f-f1d56652be2c-config-data-merged\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.032247 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-combined-ca-bundle\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.032322 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-config-data\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.034165 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-scripts\") pod \"octavia-db-sync-mjbrc\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.165039 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.916412 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-mjbrc"] Oct 04 04:44:39 crc kubenswrapper[4770]: W1004 04:44:39.929351 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod109f402b_830b_4988_9a1f_f1d56652be2c.slice/crio-e18c2e5b631a647bf19ae1b8a2399af1eef27e965f34f323030cf13aa0b4dd73 WatchSource:0}: Error finding container e18c2e5b631a647bf19ae1b8a2399af1eef27e965f34f323030cf13aa0b4dd73: Status 404 returned error can't find the container with id e18c2e5b631a647bf19ae1b8a2399af1eef27e965f34f323030cf13aa0b4dd73 Oct 04 04:44:39 crc kubenswrapper[4770]: I1004 04:44:39.992372 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-mjbrc" event={"ID":"109f402b-830b-4988-9a1f-f1d56652be2c","Type":"ContainerStarted","Data":"e18c2e5b631a647bf19ae1b8a2399af1eef27e965f34f323030cf13aa0b4dd73"} Oct 04 04:44:44 crc kubenswrapper[4770]: I1004 04:44:44.029734 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5968f659f7-n2kjd" event={"ID":"198ec713-5ae0-4e0a-ba62-433810e9e486","Type":"ContainerStarted","Data":"372178f53ef45703b352fee82ceb4917b506f100333d7179d6498a89cbe5cc6f"} Oct 04 04:44:44 crc kubenswrapper[4770]: I1004 04:44:44.031229 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-mjbrc" event={"ID":"109f402b-830b-4988-9a1f-f1d56652be2c","Type":"ContainerStarted","Data":"86a253a749e1143c875200c889641eaf602c948aea1da60992c163c8c90b94af"} Oct 04 04:44:45 crc kubenswrapper[4770]: I1004 04:44:45.044343 4770 generic.go:334] "Generic (PLEG): container finished" podID="109f402b-830b-4988-9a1f-f1d56652be2c" containerID="86a253a749e1143c875200c889641eaf602c948aea1da60992c163c8c90b94af" exitCode=0 Oct 04 04:44:45 crc kubenswrapper[4770]: I1004 04:44:45.044774 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-mjbrc" event={"ID":"109f402b-830b-4988-9a1f-f1d56652be2c","Type":"ContainerDied","Data":"86a253a749e1143c875200c889641eaf602c948aea1da60992c163c8c90b94af"} Oct 04 04:44:45 crc kubenswrapper[4770]: I1004 04:44:45.048421 4770 generic.go:334] "Generic (PLEG): container finished" podID="198ec713-5ae0-4e0a-ba62-433810e9e486" containerID="372178f53ef45703b352fee82ceb4917b506f100333d7179d6498a89cbe5cc6f" exitCode=0 Oct 04 04:44:45 crc kubenswrapper[4770]: I1004 04:44:45.048468 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5968f659f7-n2kjd" event={"ID":"198ec713-5ae0-4e0a-ba62-433810e9e486","Type":"ContainerDied","Data":"372178f53ef45703b352fee82ceb4917b506f100333d7179d6498a89cbe5cc6f"} Oct 04 04:44:46 crc kubenswrapper[4770]: I1004 04:44:46.059967 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5968f659f7-n2kjd" event={"ID":"198ec713-5ae0-4e0a-ba62-433810e9e486","Type":"ContainerStarted","Data":"ceaffe94d12c3493cbada62e25ab48203b3504414b5d657373b571c798b61161"} Oct 04 04:44:46 crc kubenswrapper[4770]: I1004 04:44:46.060344 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5968f659f7-n2kjd" event={"ID":"198ec713-5ae0-4e0a-ba62-433810e9e486","Type":"ContainerStarted","Data":"6e303cbfe89787c396e6ab93a13c106c7c81d86eba19105fbd77af20cfcacc1a"} Oct 04 04:44:46 crc kubenswrapper[4770]: I1004 04:44:46.060561 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:46 crc kubenswrapper[4770]: I1004 04:44:46.063702 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-mjbrc" event={"ID":"109f402b-830b-4988-9a1f-f1d56652be2c","Type":"ContainerStarted","Data":"685520aeef311df4260e8f8767f9d6baef1790b2fc015bca3bcbb729e1f612ad"} Oct 04 04:44:46 crc kubenswrapper[4770]: I1004 04:44:46.078685 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-5968f659f7-n2kjd" podStartSLOduration=3.403688502 podStartE2EDuration="46.078667541s" podCreationTimestamp="2025-10-04 04:44:00 +0000 UTC" firstStartedPulling="2025-10-04 04:44:00.840195853 +0000 UTC m=+6052.132205565" lastFinishedPulling="2025-10-04 04:44:43.515174882 +0000 UTC m=+6094.807184604" observedRunningTime="2025-10-04 04:44:46.076458483 +0000 UTC m=+6097.368468215" watchObservedRunningTime="2025-10-04 04:44:46.078667541 +0000 UTC m=+6097.370677263" Oct 04 04:44:46 crc kubenswrapper[4770]: I1004 04:44:46.108977 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-mjbrc" podStartSLOduration=4.522427906 podStartE2EDuration="8.108958193s" podCreationTimestamp="2025-10-04 04:44:38 +0000 UTC" firstStartedPulling="2025-10-04 04:44:39.930903544 +0000 UTC m=+6091.222913256" lastFinishedPulling="2025-10-04 04:44:43.517433831 +0000 UTC m=+6094.809443543" observedRunningTime="2025-10-04 04:44:46.104656041 +0000 UTC m=+6097.396665763" watchObservedRunningTime="2025-10-04 04:44:46.108958193 +0000 UTC m=+6097.400967905" Oct 04 04:44:47 crc kubenswrapper[4770]: I1004 04:44:47.083392 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:44:48 crc kubenswrapper[4770]: I1004 04:44:48.674350 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:44:48 crc kubenswrapper[4770]: E1004 04:44:48.675029 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:44:54 crc kubenswrapper[4770]: I1004 04:44:54.159039 4770 generic.go:334] "Generic (PLEG): container finished" podID="109f402b-830b-4988-9a1f-f1d56652be2c" containerID="685520aeef311df4260e8f8767f9d6baef1790b2fc015bca3bcbb729e1f612ad" exitCode=0 Oct 04 04:44:54 crc kubenswrapper[4770]: I1004 04:44:54.159265 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-mjbrc" event={"ID":"109f402b-830b-4988-9a1f-f1d56652be2c","Type":"ContainerDied","Data":"685520aeef311df4260e8f8767f9d6baef1790b2fc015bca3bcbb729e1f612ad"} Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.555850 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.659218 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-config-data\") pod \"109f402b-830b-4988-9a1f-f1d56652be2c\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.659313 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-combined-ca-bundle\") pod \"109f402b-830b-4988-9a1f-f1d56652be2c\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.659347 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/109f402b-830b-4988-9a1f-f1d56652be2c-config-data-merged\") pod \"109f402b-830b-4988-9a1f-f1d56652be2c\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.661633 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-scripts\") pod \"109f402b-830b-4988-9a1f-f1d56652be2c\" (UID: \"109f402b-830b-4988-9a1f-f1d56652be2c\") " Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.668489 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-config-data" (OuterVolumeSpecName: "config-data") pod "109f402b-830b-4988-9a1f-f1d56652be2c" (UID: "109f402b-830b-4988-9a1f-f1d56652be2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.680882 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-scripts" (OuterVolumeSpecName: "scripts") pod "109f402b-830b-4988-9a1f-f1d56652be2c" (UID: "109f402b-830b-4988-9a1f-f1d56652be2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.694241 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/109f402b-830b-4988-9a1f-f1d56652be2c-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "109f402b-830b-4988-9a1f-f1d56652be2c" (UID: "109f402b-830b-4988-9a1f-f1d56652be2c"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.698572 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "109f402b-830b-4988-9a1f-f1d56652be2c" (UID: "109f402b-830b-4988-9a1f-f1d56652be2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.764393 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.764420 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.764431 4770 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/109f402b-830b-4988-9a1f-f1d56652be2c-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 04 04:44:55 crc kubenswrapper[4770]: I1004 04:44:55.764439 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/109f402b-830b-4988-9a1f-f1d56652be2c-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:44:56 crc kubenswrapper[4770]: I1004 04:44:56.043851 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-9kk82"] Oct 04 04:44:56 crc kubenswrapper[4770]: I1004 04:44:56.052617 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-9kk82"] Oct 04 04:44:56 crc kubenswrapper[4770]: I1004 04:44:56.181963 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-mjbrc" event={"ID":"109f402b-830b-4988-9a1f-f1d56652be2c","Type":"ContainerDied","Data":"e18c2e5b631a647bf19ae1b8a2399af1eef27e965f34f323030cf13aa0b4dd73"} Oct 04 04:44:56 crc kubenswrapper[4770]: I1004 04:44:56.182077 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e18c2e5b631a647bf19ae1b8a2399af1eef27e965f34f323030cf13aa0b4dd73" Oct 04 04:44:56 crc kubenswrapper[4770]: I1004 04:44:56.182148 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-mjbrc" Oct 04 04:44:57 crc kubenswrapper[4770]: I1004 04:44:57.688572 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="905a1811-a6a0-4172-8d4b-1a34bf20865a" path="/var/lib/kubelet/pods/905a1811-a6a0-4172-8d4b-1a34bf20865a/volumes" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.153701 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7"] Oct 04 04:45:00 crc kubenswrapper[4770]: E1004 04:45:00.155066 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109f402b-830b-4988-9a1f-f1d56652be2c" containerName="init" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.155089 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="109f402b-830b-4988-9a1f-f1d56652be2c" containerName="init" Oct 04 04:45:00 crc kubenswrapper[4770]: E1004 04:45:00.155134 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109f402b-830b-4988-9a1f-f1d56652be2c" containerName="octavia-db-sync" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.155141 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="109f402b-830b-4988-9a1f-f1d56652be2c" containerName="octavia-db-sync" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.155405 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="109f402b-830b-4988-9a1f-f1d56652be2c" containerName="octavia-db-sync" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.156439 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.158765 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.159468 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.165131 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7"] Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.259621 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnwrx\" (UniqueName: \"kubernetes.io/projected/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-kube-api-access-xnwrx\") pod \"collect-profiles-29325885-gvqx7\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.259683 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-secret-volume\") pod \"collect-profiles-29325885-gvqx7\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.259978 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-config-volume\") pod \"collect-profiles-29325885-gvqx7\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.363147 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-secret-volume\") pod \"collect-profiles-29325885-gvqx7\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.363289 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-config-volume\") pod \"collect-profiles-29325885-gvqx7\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.363504 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnwrx\" (UniqueName: \"kubernetes.io/projected/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-kube-api-access-xnwrx\") pod \"collect-profiles-29325885-gvqx7\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.364556 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-config-volume\") pod \"collect-profiles-29325885-gvqx7\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.373695 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-secret-volume\") pod \"collect-profiles-29325885-gvqx7\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.384963 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnwrx\" (UniqueName: \"kubernetes.io/projected/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-kube-api-access-xnwrx\") pod \"collect-profiles-29325885-gvqx7\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.489743 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.677198 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:45:00 crc kubenswrapper[4770]: E1004 04:45:00.677793 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:45:00 crc kubenswrapper[4770]: I1004 04:45:00.973663 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7"] Oct 04 04:45:01 crc kubenswrapper[4770]: I1004 04:45:01.259285 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" event={"ID":"51eba811-6f6a-4055-9f07-fdcf4ddc9ced","Type":"ContainerStarted","Data":"a10148d1aaaee1d1e1311cba49494469b41b4309448843a0c88d58fd11e18306"} Oct 04 04:45:01 crc kubenswrapper[4770]: I1004 04:45:01.259618 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" event={"ID":"51eba811-6f6a-4055-9f07-fdcf4ddc9ced","Type":"ContainerStarted","Data":"640c15c49c98ac8dc359fddf129241947042956f5ca7e4a6f69583159df7476b"} Oct 04 04:45:01 crc kubenswrapper[4770]: I1004 04:45:01.281982 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" podStartSLOduration=1.281950358 podStartE2EDuration="1.281950358s" podCreationTimestamp="2025-10-04 04:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:45:01.278393895 +0000 UTC m=+6112.570403607" watchObservedRunningTime="2025-10-04 04:45:01.281950358 +0000 UTC m=+6112.573960070" Oct 04 04:45:02 crc kubenswrapper[4770]: I1004 04:45:02.268433 4770 generic.go:334] "Generic (PLEG): container finished" podID="51eba811-6f6a-4055-9f07-fdcf4ddc9ced" containerID="a10148d1aaaee1d1e1311cba49494469b41b4309448843a0c88d58fd11e18306" exitCode=0 Oct 04 04:45:02 crc kubenswrapper[4770]: I1004 04:45:02.268527 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" event={"ID":"51eba811-6f6a-4055-9f07-fdcf4ddc9ced","Type":"ContainerDied","Data":"a10148d1aaaee1d1e1311cba49494469b41b4309448843a0c88d58fd11e18306"} Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.680382 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.847287 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-config-volume\") pod \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.847392 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnwrx\" (UniqueName: \"kubernetes.io/projected/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-kube-api-access-xnwrx\") pod \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.847432 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-secret-volume\") pod \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\" (UID: \"51eba811-6f6a-4055-9f07-fdcf4ddc9ced\") " Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.848170 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-config-volume" (OuterVolumeSpecName: "config-volume") pod "51eba811-6f6a-4055-9f07-fdcf4ddc9ced" (UID: "51eba811-6f6a-4055-9f07-fdcf4ddc9ced"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.853038 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "51eba811-6f6a-4055-9f07-fdcf4ddc9ced" (UID: "51eba811-6f6a-4055-9f07-fdcf4ddc9ced"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.853095 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-kube-api-access-xnwrx" (OuterVolumeSpecName: "kube-api-access-xnwrx") pod "51eba811-6f6a-4055-9f07-fdcf4ddc9ced" (UID: "51eba811-6f6a-4055-9f07-fdcf4ddc9ced"). InnerVolumeSpecName "kube-api-access-xnwrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.950197 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnwrx\" (UniqueName: \"kubernetes.io/projected/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-kube-api-access-xnwrx\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.950231 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:03 crc kubenswrapper[4770]: I1004 04:45:03.950242 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51eba811-6f6a-4055-9f07-fdcf4ddc9ced-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:04 crc kubenswrapper[4770]: I1004 04:45:04.292048 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" event={"ID":"51eba811-6f6a-4055-9f07-fdcf4ddc9ced","Type":"ContainerDied","Data":"640c15c49c98ac8dc359fddf129241947042956f5ca7e4a6f69583159df7476b"} Oct 04 04:45:04 crc kubenswrapper[4770]: I1004 04:45:04.292099 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="640c15c49c98ac8dc359fddf129241947042956f5ca7e4a6f69583159df7476b" Oct 04 04:45:04 crc kubenswrapper[4770]: I1004 04:45:04.292112 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7" Oct 04 04:45:04 crc kubenswrapper[4770]: I1004 04:45:04.419316 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:45:04 crc kubenswrapper[4770]: I1004 04:45:04.552281 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-5968f659f7-n2kjd" Oct 04 04:45:04 crc kubenswrapper[4770]: I1004 04:45:04.759550 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4"] Oct 04 04:45:04 crc kubenswrapper[4770]: I1004 04:45:04.767795 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-t68f4"] Oct 04 04:45:05 crc kubenswrapper[4770]: I1004 04:45:05.844150 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3463bd44-4352-4f30-9ebd-ac8667f4db7f" path="/var/lib/kubelet/pods/3463bd44-4352-4f30-9ebd-ac8667f4db7f/volumes" Oct 04 04:45:06 crc kubenswrapper[4770]: I1004 04:45:06.032513 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-685f-account-create-7zfxz"] Oct 04 04:45:06 crc kubenswrapper[4770]: I1004 04:45:06.053547 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-685f-account-create-7zfxz"] Oct 04 04:45:07 crc kubenswrapper[4770]: I1004 04:45:07.685426 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0846dbf-391e-4431-923e-1bf4d138be4b" path="/var/lib/kubelet/pods/e0846dbf-391e-4431-923e-1bf4d138be4b/volumes" Oct 04 04:45:11 crc kubenswrapper[4770]: I1004 04:45:11.673796 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:45:11 crc kubenswrapper[4770]: E1004 04:45:11.674382 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:45:15 crc kubenswrapper[4770]: I1004 04:45:15.022822 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-t5797"] Oct 04 04:45:15 crc kubenswrapper[4770]: I1004 04:45:15.037084 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-t5797"] Oct 04 04:45:15 crc kubenswrapper[4770]: I1004 04:45:15.686103 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bf96ca3-0b4a-489a-864a-d0299e1db7cc" path="/var/lib/kubelet/pods/7bf96ca3-0b4a-489a-864a-d0299e1db7cc/volumes" Oct 04 04:45:24 crc kubenswrapper[4770]: I1004 04:45:24.680110 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:45:24 crc kubenswrapper[4770]: E1004 04:45:24.681978 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.612793 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-jxh7k"] Oct 04 04:45:25 crc kubenswrapper[4770]: E1004 04:45:25.613309 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51eba811-6f6a-4055-9f07-fdcf4ddc9ced" containerName="collect-profiles" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.613333 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="51eba811-6f6a-4055-9f07-fdcf4ddc9ced" containerName="collect-profiles" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.613612 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="51eba811-6f6a-4055-9f07-fdcf4ddc9ced" containerName="collect-profiles" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.614807 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.617181 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.617233 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.617256 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.630232 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-jxh7k"] Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.710017 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c388dfff-1c4e-4b28-9998-edea3809767b-config-data\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.710167 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c388dfff-1c4e-4b28-9998-edea3809767b-config-data-merged\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.710353 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/c388dfff-1c4e-4b28-9998-edea3809767b-hm-ports\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.710854 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c388dfff-1c4e-4b28-9998-edea3809767b-scripts\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.812592 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/c388dfff-1c4e-4b28-9998-edea3809767b-hm-ports\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.812723 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c388dfff-1c4e-4b28-9998-edea3809767b-scripts\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.812782 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c388dfff-1c4e-4b28-9998-edea3809767b-config-data\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.812830 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c388dfff-1c4e-4b28-9998-edea3809767b-config-data-merged\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.813455 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c388dfff-1c4e-4b28-9998-edea3809767b-config-data-merged\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.813791 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/c388dfff-1c4e-4b28-9998-edea3809767b-hm-ports\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.823159 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c388dfff-1c4e-4b28-9998-edea3809767b-config-data\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.827945 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c388dfff-1c4e-4b28-9998-edea3809767b-scripts\") pod \"octavia-rsyslog-jxh7k\" (UID: \"c388dfff-1c4e-4b28-9998-edea3809767b\") " pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:25 crc kubenswrapper[4770]: I1004 04:45:25.940180 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.513615 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-jxh7k"] Oct 04 04:45:26 crc kubenswrapper[4770]: W1004 04:45:26.521970 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc388dfff_1c4e_4b28_9998_edea3809767b.slice/crio-c05314492c5becbba70fd9d6a9247065a61acd953b91225e824d9e576b2165cf WatchSource:0}: Error finding container c05314492c5becbba70fd9d6a9247065a61acd953b91225e824d9e576b2165cf: Status 404 returned error can't find the container with id c05314492c5becbba70fd9d6a9247065a61acd953b91225e824d9e576b2165cf Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.563334 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-jxh7k" event={"ID":"c388dfff-1c4e-4b28-9998-edea3809767b","Type":"ContainerStarted","Data":"c05314492c5becbba70fd9d6a9247065a61acd953b91225e824d9e576b2165cf"} Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.589260 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-678599687f-9j55c"] Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.591000 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.594915 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.601563 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-9j55c"] Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.628277 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/93e6caa9-b2cd-48e5-b332-5217fb8f110b-amphora-image\") pod \"octavia-image-upload-678599687f-9j55c\" (UID: \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\") " pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.628333 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/93e6caa9-b2cd-48e5-b332-5217fb8f110b-httpd-config\") pod \"octavia-image-upload-678599687f-9j55c\" (UID: \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\") " pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.730843 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/93e6caa9-b2cd-48e5-b332-5217fb8f110b-amphora-image\") pod \"octavia-image-upload-678599687f-9j55c\" (UID: \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\") " pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.731537 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/93e6caa9-b2cd-48e5-b332-5217fb8f110b-httpd-config\") pod \"octavia-image-upload-678599687f-9j55c\" (UID: \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\") " pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.731463 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/93e6caa9-b2cd-48e5-b332-5217fb8f110b-amphora-image\") pod \"octavia-image-upload-678599687f-9j55c\" (UID: \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\") " pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.737202 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/93e6caa9-b2cd-48e5-b332-5217fb8f110b-httpd-config\") pod \"octavia-image-upload-678599687f-9j55c\" (UID: \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\") " pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:45:26 crc kubenswrapper[4770]: I1004 04:45:26.918614 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:45:27 crc kubenswrapper[4770]: I1004 04:45:27.247454 4770 scope.go:117] "RemoveContainer" containerID="adbc6ea3b629a932317e2bc40100cb81459d77694b33e6802b7d653870f344ec" Oct 04 04:45:27 crc kubenswrapper[4770]: I1004 04:45:27.285224 4770 scope.go:117] "RemoveContainer" containerID="44bbe576224173f3afb1a8b498626bffcd059b8b3391a2d339cfb57981fc7ce0" Oct 04 04:45:27 crc kubenswrapper[4770]: I1004 04:45:27.365903 4770 scope.go:117] "RemoveContainer" containerID="ff3bd21e1fd61fb6741aa77b9ef6bf4e849ceb0e4089b0a8419e3d8671ca3f43" Oct 04 04:45:27 crc kubenswrapper[4770]: I1004 04:45:27.405238 4770 scope.go:117] "RemoveContainer" containerID="5fc058ab5405739055c9607f76969dd47f409c9f692576783792c47157490033" Oct 04 04:45:27 crc kubenswrapper[4770]: I1004 04:45:27.405508 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-9j55c"] Oct 04 04:45:27 crc kubenswrapper[4770]: I1004 04:45:27.434343 4770 scope.go:117] "RemoveContainer" containerID="6a51f95acd2f96ec620189e82a3590543e173291f1fb0f675b7e4878664cb850" Oct 04 04:45:27 crc kubenswrapper[4770]: W1004 04:45:27.434502 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93e6caa9_b2cd_48e5_b332_5217fb8f110b.slice/crio-dac8d9662538f231859e0e6f2c7a83de23067f2002387313b74ff72fb09ee597 WatchSource:0}: Error finding container dac8d9662538f231859e0e6f2c7a83de23067f2002387313b74ff72fb09ee597: Status 404 returned error can't find the container with id dac8d9662538f231859e0e6f2c7a83de23067f2002387313b74ff72fb09ee597 Oct 04 04:45:27 crc kubenswrapper[4770]: I1004 04:45:27.467521 4770 scope.go:117] "RemoveContainer" containerID="dfb8fc47ee82654121d34d69f42ec706252093c7c3ae63bbb1f7581212e8ab88" Oct 04 04:45:27 crc kubenswrapper[4770]: I1004 04:45:27.577303 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-9j55c" event={"ID":"93e6caa9-b2cd-48e5-b332-5217fb8f110b","Type":"ContainerStarted","Data":"dac8d9662538f231859e0e6f2c7a83de23067f2002387313b74ff72fb09ee597"} Oct 04 04:45:31 crc kubenswrapper[4770]: I1004 04:45:31.620283 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-jxh7k" event={"ID":"c388dfff-1c4e-4b28-9998-edea3809767b","Type":"ContainerStarted","Data":"3a6abceed57daf9961abbf11539bf595f787f597e6f918ee3c96d4d7d68802f4"} Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.329738 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-47zrr"] Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.332367 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.336691 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.336857 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.336993 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.355391 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-combined-ca-bundle\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.355726 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-config-data-merged\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.355756 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-config-data\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.355801 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-hm-ports\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.355823 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-scripts\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.355874 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-amphora-certs\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.362867 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-47zrr"] Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.458633 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-amphora-certs\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.458804 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-combined-ca-bundle\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.458850 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-config-data-merged\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.459452 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-config-data-merged\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.459518 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-config-data\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.459946 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-hm-ports\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.460104 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-scripts\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.461967 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-hm-ports\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.469547 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-combined-ca-bundle\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.469629 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-config-data\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.470838 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-amphora-certs\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.470926 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67bf6fd1-b5eb-4c67-8210-0839f7cdb14b-scripts\") pod \"octavia-healthmanager-47zrr\" (UID: \"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b\") " pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.640317 4770 generic.go:334] "Generic (PLEG): container finished" podID="c388dfff-1c4e-4b28-9998-edea3809767b" containerID="3a6abceed57daf9961abbf11539bf595f787f597e6f918ee3c96d4d7d68802f4" exitCode=0 Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.640365 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-jxh7k" event={"ID":"c388dfff-1c4e-4b28-9998-edea3809767b","Type":"ContainerDied","Data":"3a6abceed57daf9961abbf11539bf595f787f597e6f918ee3c96d4d7d68802f4"} Oct 04 04:45:33 crc kubenswrapper[4770]: I1004 04:45:33.659619 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:34 crc kubenswrapper[4770]: I1004 04:45:34.375714 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-47zrr"] Oct 04 04:45:34 crc kubenswrapper[4770]: W1004 04:45:34.389752 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67bf6fd1_b5eb_4c67_8210_0839f7cdb14b.slice/crio-1e82f0439c0d3fdfd9a74e33d2b7d77d34e2e4d32b731ea63300b0b1f0304167 WatchSource:0}: Error finding container 1e82f0439c0d3fdfd9a74e33d2b7d77d34e2e4d32b731ea63300b0b1f0304167: Status 404 returned error can't find the container with id 1e82f0439c0d3fdfd9a74e33d2b7d77d34e2e4d32b731ea63300b0b1f0304167 Oct 04 04:45:34 crc kubenswrapper[4770]: I1004 04:45:34.651225 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-47zrr" event={"ID":"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b","Type":"ContainerStarted","Data":"1e82f0439c0d3fdfd9a74e33d2b7d77d34e2e4d32b731ea63300b0b1f0304167"} Oct 04 04:45:36 crc kubenswrapper[4770]: I1004 04:45:36.673854 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:45:36 crc kubenswrapper[4770]: E1004 04:45:36.674530 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.795238 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-wx6tq"] Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.797245 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.799282 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.799534 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.816429 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-wx6tq"] Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.856732 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-amphora-certs\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.856783 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/aadc601e-a24b-4258-91c5-c1f703ed1973-hm-ports\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.856922 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-config-data\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.857180 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-combined-ca-bundle\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.857292 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-scripts\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.857375 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aadc601e-a24b-4258-91c5-c1f703ed1973-config-data-merged\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.959078 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-scripts\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.959408 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aadc601e-a24b-4258-91c5-c1f703ed1973-config-data-merged\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.959530 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-amphora-certs\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.959649 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/aadc601e-a24b-4258-91c5-c1f703ed1973-hm-ports\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.959822 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aadc601e-a24b-4258-91c5-c1f703ed1973-config-data-merged\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.960632 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/aadc601e-a24b-4258-91c5-c1f703ed1973-hm-ports\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.961046 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-config-data\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.961661 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-combined-ca-bundle\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.966107 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-combined-ca-bundle\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.967112 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-config-data\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.968636 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-amphora-certs\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:37 crc kubenswrapper[4770]: I1004 04:45:37.980028 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aadc601e-a24b-4258-91c5-c1f703ed1973-scripts\") pod \"octavia-housekeeping-wx6tq\" (UID: \"aadc601e-a24b-4258-91c5-c1f703ed1973\") " pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:38 crc kubenswrapper[4770]: I1004 04:45:38.122156 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.689135 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-d69sh"] Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.691077 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.692910 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.695096 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.704602 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-d69sh"] Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.721575 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-config-data\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.721653 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/11a80330-ab5a-4f00-b3b4-341d6d23a268-hm-ports\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.721682 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-combined-ca-bundle\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.721712 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/11a80330-ab5a-4f00-b3b4-341d6d23a268-config-data-merged\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.721799 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-scripts\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.721824 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-amphora-certs\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.824409 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-config-data\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.824495 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/11a80330-ab5a-4f00-b3b4-341d6d23a268-hm-ports\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.824527 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-combined-ca-bundle\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.824566 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/11a80330-ab5a-4f00-b3b4-341d6d23a268-config-data-merged\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.824658 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-scripts\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.824685 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-amphora-certs\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.825401 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/11a80330-ab5a-4f00-b3b4-341d6d23a268-config-data-merged\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.825785 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/11a80330-ab5a-4f00-b3b4-341d6d23a268-hm-ports\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.831775 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-scripts\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.832045 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-config-data\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.832156 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-amphora-certs\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:40 crc kubenswrapper[4770]: I1004 04:45:40.833788 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a80330-ab5a-4f00-b3b4-341d6d23a268-combined-ca-bundle\") pod \"octavia-worker-d69sh\" (UID: \"11a80330-ab5a-4f00-b3b4-341d6d23a268\") " pod="openstack/octavia-worker-d69sh" Oct 04 04:45:41 crc kubenswrapper[4770]: I1004 04:45:41.023353 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-d69sh" Oct 04 04:45:45 crc kubenswrapper[4770]: I1004 04:45:45.038425 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-dql4d"] Oct 04 04:45:45 crc kubenswrapper[4770]: I1004 04:45:45.048887 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-dql4d"] Oct 04 04:45:45 crc kubenswrapper[4770]: I1004 04:45:45.688072 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c53d120-d264-4191-884d-e377d0ffbb2e" path="/var/lib/kubelet/pods/5c53d120-d264-4191-884d-e377d0ffbb2e/volumes" Oct 04 04:45:47 crc kubenswrapper[4770]: I1004 04:45:47.674184 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:45:47 crc kubenswrapper[4770]: E1004 04:45:47.675158 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:45:49 crc kubenswrapper[4770]: E1004 04:45:49.741120 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/gthiemonge/octavia-amphora-image:latest" Oct 04 04:45:49 crc kubenswrapper[4770]: E1004 04:45:49.742106 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/gthiemonge/octavia-amphora-image,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DEST_DIR,Value:/usr/local/apache2/htdocs,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:amphora-image,ReadOnly:false,MountPath:/usr/local/apache2/htdocs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-image-upload-678599687f-9j55c_openstack(93e6caa9-b2cd-48e5-b332-5217fb8f110b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:45:49 crc kubenswrapper[4770]: E1004 04:45:49.745038 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/octavia-image-upload-678599687f-9j55c" podUID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" Oct 04 04:45:49 crc kubenswrapper[4770]: I1004 04:45:49.833387 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-47zrr" event={"ID":"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b","Type":"ContainerStarted","Data":"9edb838d4e1c6a71658871282181997bbd1567ebcd70e37decfc8f36ee9cc4a6"} Oct 04 04:45:49 crc kubenswrapper[4770]: E1004 04:45:49.834753 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/gthiemonge/octavia-amphora-image\\\"\"" pod="openstack/octavia-image-upload-678599687f-9j55c" podUID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" Oct 04 04:45:49 crc kubenswrapper[4770]: I1004 04:45:49.889752 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-wx6tq"] Oct 04 04:45:49 crc kubenswrapper[4770]: W1004 04:45:49.891440 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaadc601e_a24b_4258_91c5_c1f703ed1973.slice/crio-09d6850156d080bfc1fe5eda6368b8e4ece9d9e0e3ceccbcee8ebac737a066de WatchSource:0}: Error finding container 09d6850156d080bfc1fe5eda6368b8e4ece9d9e0e3ceccbcee8ebac737a066de: Status 404 returned error can't find the container with id 09d6850156d080bfc1fe5eda6368b8e4ece9d9e0e3ceccbcee8ebac737a066de Oct 04 04:45:50 crc kubenswrapper[4770]: W1004 04:45:50.167684 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11a80330_ab5a_4f00_b3b4_341d6d23a268.slice/crio-3a7ccc8ed1ce7f373b9280b2841a848d2ac26f273ff67d5fa052103d4f5fc426 WatchSource:0}: Error finding container 3a7ccc8ed1ce7f373b9280b2841a848d2ac26f273ff67d5fa052103d4f5fc426: Status 404 returned error can't find the container with id 3a7ccc8ed1ce7f373b9280b2841a848d2ac26f273ff67d5fa052103d4f5fc426 Oct 04 04:45:50 crc kubenswrapper[4770]: I1004 04:45:50.180089 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-d69sh"] Oct 04 04:45:50 crc kubenswrapper[4770]: I1004 04:45:50.850039 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-jxh7k" event={"ID":"c388dfff-1c4e-4b28-9998-edea3809767b","Type":"ContainerStarted","Data":"4ea3c9c936be1c02b2ce002236e7ad58861be49e2ac8714a87a6380e266d9958"} Oct 04 04:45:50 crc kubenswrapper[4770]: I1004 04:45:50.850337 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:50 crc kubenswrapper[4770]: I1004 04:45:50.852276 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-wx6tq" event={"ID":"aadc601e-a24b-4258-91c5-c1f703ed1973","Type":"ContainerStarted","Data":"09d6850156d080bfc1fe5eda6368b8e4ece9d9e0e3ceccbcee8ebac737a066de"} Oct 04 04:45:50 crc kubenswrapper[4770]: I1004 04:45:50.854831 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-d69sh" event={"ID":"11a80330-ab5a-4f00-b3b4-341d6d23a268","Type":"ContainerStarted","Data":"3a7ccc8ed1ce7f373b9280b2841a848d2ac26f273ff67d5fa052103d4f5fc426"} Oct 04 04:45:50 crc kubenswrapper[4770]: I1004 04:45:50.857746 4770 generic.go:334] "Generic (PLEG): container finished" podID="67bf6fd1-b5eb-4c67-8210-0839f7cdb14b" containerID="9edb838d4e1c6a71658871282181997bbd1567ebcd70e37decfc8f36ee9cc4a6" exitCode=0 Oct 04 04:45:50 crc kubenswrapper[4770]: I1004 04:45:50.857777 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-47zrr" event={"ID":"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b","Type":"ContainerDied","Data":"9edb838d4e1c6a71658871282181997bbd1567ebcd70e37decfc8f36ee9cc4a6"} Oct 04 04:45:50 crc kubenswrapper[4770]: I1004 04:45:50.884886 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-jxh7k" podStartSLOduration=2.899435125 podStartE2EDuration="25.884868397s" podCreationTimestamp="2025-10-04 04:45:25 +0000 UTC" firstStartedPulling="2025-10-04 04:45:26.525168211 +0000 UTC m=+6137.817177933" lastFinishedPulling="2025-10-04 04:45:49.510601493 +0000 UTC m=+6160.802611205" observedRunningTime="2025-10-04 04:45:50.874089725 +0000 UTC m=+6162.166099437" watchObservedRunningTime="2025-10-04 04:45:50.884868397 +0000 UTC m=+6162.176878109" Oct 04 04:45:51 crc kubenswrapper[4770]: I1004 04:45:51.931873 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-47zrr" event={"ID":"67bf6fd1-b5eb-4c67-8210-0839f7cdb14b","Type":"ContainerStarted","Data":"0dd322650dfb03c705d6d1b5cb81cc56951bf31b9f413cbd41a05f1ca126d522"} Oct 04 04:45:51 crc kubenswrapper[4770]: I1004 04:45:51.932361 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:45:52 crc kubenswrapper[4770]: I1004 04:45:52.941669 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-wx6tq" event={"ID":"aadc601e-a24b-4258-91c5-c1f703ed1973","Type":"ContainerStarted","Data":"de11f4b336d90a506458c1e3baea0cbe5ee0dd459902fb3c32a2880e207407b5"} Oct 04 04:45:52 crc kubenswrapper[4770]: I1004 04:45:52.946673 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-d69sh" event={"ID":"11a80330-ab5a-4f00-b3b4-341d6d23a268","Type":"ContainerStarted","Data":"0fb5d09e853c3e985da6ff57aaa9f6085987b6ea3ba7ef76bd2ab94831c06ce9"} Oct 04 04:45:52 crc kubenswrapper[4770]: I1004 04:45:52.972778 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-47zrr" podStartSLOduration=19.972752297 podStartE2EDuration="19.972752297s" podCreationTimestamp="2025-10-04 04:45:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:45:51.971513569 +0000 UTC m=+6163.263523291" watchObservedRunningTime="2025-10-04 04:45:52.972752297 +0000 UTC m=+6164.264762029" Oct 04 04:45:53 crc kubenswrapper[4770]: I1004 04:45:53.959980 4770 generic.go:334] "Generic (PLEG): container finished" podID="aadc601e-a24b-4258-91c5-c1f703ed1973" containerID="de11f4b336d90a506458c1e3baea0cbe5ee0dd459902fb3c32a2880e207407b5" exitCode=0 Oct 04 04:45:53 crc kubenswrapper[4770]: I1004 04:45:53.960248 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-wx6tq" event={"ID":"aadc601e-a24b-4258-91c5-c1f703ed1973","Type":"ContainerDied","Data":"de11f4b336d90a506458c1e3baea0cbe5ee0dd459902fb3c32a2880e207407b5"} Oct 04 04:45:53 crc kubenswrapper[4770]: I1004 04:45:53.962841 4770 generic.go:334] "Generic (PLEG): container finished" podID="11a80330-ab5a-4f00-b3b4-341d6d23a268" containerID="0fb5d09e853c3e985da6ff57aaa9f6085987b6ea3ba7ef76bd2ab94831c06ce9" exitCode=0 Oct 04 04:45:53 crc kubenswrapper[4770]: I1004 04:45:53.964343 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-d69sh" event={"ID":"11a80330-ab5a-4f00-b3b4-341d6d23a268","Type":"ContainerDied","Data":"0fb5d09e853c3e985da6ff57aaa9f6085987b6ea3ba7ef76bd2ab94831c06ce9"} Oct 04 04:45:54 crc kubenswrapper[4770]: I1004 04:45:54.985204 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-wx6tq" event={"ID":"aadc601e-a24b-4258-91c5-c1f703ed1973","Type":"ContainerStarted","Data":"f48b08f1820ffe3b3dc968737946fed99f8cda281858ca052d806f675c2580f5"} Oct 04 04:45:54 crc kubenswrapper[4770]: I1004 04:45:54.985714 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:45:54 crc kubenswrapper[4770]: I1004 04:45:54.991955 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-d69sh" event={"ID":"11a80330-ab5a-4f00-b3b4-341d6d23a268","Type":"ContainerStarted","Data":"e19b2d5e7a8fec33a48c15408e71ece47317d0767bbbaa37ff61fe00ef552f0b"} Oct 04 04:45:54 crc kubenswrapper[4770]: I1004 04:45:54.992896 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-d69sh" Oct 04 04:45:55 crc kubenswrapper[4770]: I1004 04:45:55.009964 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-wx6tq" podStartSLOduration=15.548267024 podStartE2EDuration="18.00994032s" podCreationTimestamp="2025-10-04 04:45:37 +0000 UTC" firstStartedPulling="2025-10-04 04:45:49.900690756 +0000 UTC m=+6161.192700468" lastFinishedPulling="2025-10-04 04:45:52.362364012 +0000 UTC m=+6163.654373764" observedRunningTime="2025-10-04 04:45:55.005421451 +0000 UTC m=+6166.297431163" watchObservedRunningTime="2025-10-04 04:45:55.00994032 +0000 UTC m=+6166.301950032" Oct 04 04:45:55 crc kubenswrapper[4770]: I1004 04:45:55.028802 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7e7a-account-create-5nl84"] Oct 04 04:45:55 crc kubenswrapper[4770]: I1004 04:45:55.036845 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7e7a-account-create-5nl84"] Oct 04 04:45:55 crc kubenswrapper[4770]: I1004 04:45:55.045128 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-d69sh" podStartSLOduration=12.857652906 podStartE2EDuration="15.04510668s" podCreationTimestamp="2025-10-04 04:45:40 +0000 UTC" firstStartedPulling="2025-10-04 04:45:50.170109082 +0000 UTC m=+6161.462118784" lastFinishedPulling="2025-10-04 04:45:52.357562816 +0000 UTC m=+6163.649572558" observedRunningTime="2025-10-04 04:45:55.02981702 +0000 UTC m=+6166.321826732" watchObservedRunningTime="2025-10-04 04:45:55.04510668 +0000 UTC m=+6166.337116392" Oct 04 04:45:55 crc kubenswrapper[4770]: I1004 04:45:55.689844 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c9dba9-f1c3-4402-8c94-39c3c36807f7" path="/var/lib/kubelet/pods/12c9dba9-f1c3-4402-8c94-39c3c36807f7/volumes" Oct 04 04:45:55 crc kubenswrapper[4770]: I1004 04:45:55.971662 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-jxh7k" Oct 04 04:45:58 crc kubenswrapper[4770]: I1004 04:45:58.674353 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:45:58 crc kubenswrapper[4770]: E1004 04:45:58.675044 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:46:03 crc kubenswrapper[4770]: I1004 04:46:03.062918 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-cmdwr"] Oct 04 04:46:03 crc kubenswrapper[4770]: I1004 04:46:03.076563 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-cmdwr"] Oct 04 04:46:03 crc kubenswrapper[4770]: I1004 04:46:03.690708 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aab0c547-9737-4e63-a6d4-fba02c488986" path="/var/lib/kubelet/pods/aab0c547-9737-4e63-a6d4-fba02c488986/volumes" Oct 04 04:46:03 crc kubenswrapper[4770]: I1004 04:46:03.694790 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-47zrr" Oct 04 04:46:04 crc kubenswrapper[4770]: I1004 04:46:04.087557 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-9j55c" event={"ID":"93e6caa9-b2cd-48e5-b332-5217fb8f110b","Type":"ContainerStarted","Data":"3bb4d9a5030e54ef869fc96c9b5ccaa9f4d533b795f36d996ee571280fa6ff27"} Oct 04 04:46:08 crc kubenswrapper[4770]: I1004 04:46:08.151347 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-wx6tq" Oct 04 04:46:09 crc kubenswrapper[4770]: I1004 04:46:09.134815 4770 generic.go:334] "Generic (PLEG): container finished" podID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" containerID="3bb4d9a5030e54ef869fc96c9b5ccaa9f4d533b795f36d996ee571280fa6ff27" exitCode=0 Oct 04 04:46:09 crc kubenswrapper[4770]: I1004 04:46:09.134866 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-9j55c" event={"ID":"93e6caa9-b2cd-48e5-b332-5217fb8f110b","Type":"ContainerDied","Data":"3bb4d9a5030e54ef869fc96c9b5ccaa9f4d533b795f36d996ee571280fa6ff27"} Oct 04 04:46:09 crc kubenswrapper[4770]: I1004 04:46:09.682221 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:46:09 crc kubenswrapper[4770]: E1004 04:46:09.683050 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:46:11 crc kubenswrapper[4770]: I1004 04:46:11.058157 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-d69sh" Oct 04 04:46:11 crc kubenswrapper[4770]: I1004 04:46:11.156511 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-9j55c" event={"ID":"93e6caa9-b2cd-48e5-b332-5217fb8f110b","Type":"ContainerStarted","Data":"90286d99281c7f6636804c01778b3f82d6bd5d21972ad23fde7098c2aa1247d1"} Oct 04 04:46:11 crc kubenswrapper[4770]: I1004 04:46:11.177438 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-678599687f-9j55c" podStartSLOduration=9.002701125 podStartE2EDuration="45.177411725s" podCreationTimestamp="2025-10-04 04:45:26 +0000 UTC" firstStartedPulling="2025-10-04 04:45:27.436223601 +0000 UTC m=+6138.728233313" lastFinishedPulling="2025-10-04 04:46:03.610934211 +0000 UTC m=+6174.902943913" observedRunningTime="2025-10-04 04:46:11.170057922 +0000 UTC m=+6182.462067634" watchObservedRunningTime="2025-10-04 04:46:11.177411725 +0000 UTC m=+6182.469421457" Oct 04 04:46:23 crc kubenswrapper[4770]: I1004 04:46:23.673478 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:46:23 crc kubenswrapper[4770]: E1004 04:46:23.675226 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:46:27 crc kubenswrapper[4770]: I1004 04:46:27.602775 4770 scope.go:117] "RemoveContainer" containerID="fa5102de2ebaabe0371080b98993396ba4f3ba23ebe9e809050caafeceec55a3" Oct 04 04:46:27 crc kubenswrapper[4770]: I1004 04:46:27.651857 4770 scope.go:117] "RemoveContainer" containerID="a0a1678c9de94d202465457eaf1d439d1611f76c3acdb960b78874809b6cff85" Oct 04 04:46:27 crc kubenswrapper[4770]: I1004 04:46:27.711408 4770 scope.go:117] "RemoveContainer" containerID="42dab1f26648852abc8e6b4050123a7f8b38fc0729b4bdfeef04dfa7472811c3" Oct 04 04:46:30 crc kubenswrapper[4770]: I1004 04:46:30.582409 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-678599687f-9j55c"] Oct 04 04:46:30 crc kubenswrapper[4770]: I1004 04:46:30.583274 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-678599687f-9j55c" podUID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" containerName="octavia-amphora-httpd" containerID="cri-o://90286d99281c7f6636804c01778b3f82d6bd5d21972ad23fde7098c2aa1247d1" gracePeriod=30 Oct 04 04:46:31 crc kubenswrapper[4770]: I1004 04:46:31.359487 4770 generic.go:334] "Generic (PLEG): container finished" podID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" containerID="90286d99281c7f6636804c01778b3f82d6bd5d21972ad23fde7098c2aa1247d1" exitCode=0 Oct 04 04:46:31 crc kubenswrapper[4770]: I1004 04:46:31.359652 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-9j55c" event={"ID":"93e6caa9-b2cd-48e5-b332-5217fb8f110b","Type":"ContainerDied","Data":"90286d99281c7f6636804c01778b3f82d6bd5d21972ad23fde7098c2aa1247d1"} Oct 04 04:46:31 crc kubenswrapper[4770]: I1004 04:46:31.858115 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:46:31 crc kubenswrapper[4770]: I1004 04:46:31.938917 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/93e6caa9-b2cd-48e5-b332-5217fb8f110b-httpd-config\") pod \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\" (UID: \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\") " Oct 04 04:46:31 crc kubenswrapper[4770]: I1004 04:46:31.939102 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/93e6caa9-b2cd-48e5-b332-5217fb8f110b-amphora-image\") pod \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\" (UID: \"93e6caa9-b2cd-48e5-b332-5217fb8f110b\") " Oct 04 04:46:31 crc kubenswrapper[4770]: I1004 04:46:31.971670 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e6caa9-b2cd-48e5-b332-5217fb8f110b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "93e6caa9-b2cd-48e5-b332-5217fb8f110b" (UID: "93e6caa9-b2cd-48e5-b332-5217fb8f110b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:46:32 crc kubenswrapper[4770]: I1004 04:46:32.007787 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93e6caa9-b2cd-48e5-b332-5217fb8f110b-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "93e6caa9-b2cd-48e5-b332-5217fb8f110b" (UID: "93e6caa9-b2cd-48e5-b332-5217fb8f110b"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:46:32 crc kubenswrapper[4770]: I1004 04:46:32.041110 4770 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/93e6caa9-b2cd-48e5-b332-5217fb8f110b-amphora-image\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:32 crc kubenswrapper[4770]: I1004 04:46:32.041179 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/93e6caa9-b2cd-48e5-b332-5217fb8f110b-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:32 crc kubenswrapper[4770]: I1004 04:46:32.372322 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-9j55c" event={"ID":"93e6caa9-b2cd-48e5-b332-5217fb8f110b","Type":"ContainerDied","Data":"dac8d9662538f231859e0e6f2c7a83de23067f2002387313b74ff72fb09ee597"} Oct 04 04:46:32 crc kubenswrapper[4770]: I1004 04:46:32.372394 4770 scope.go:117] "RemoveContainer" containerID="90286d99281c7f6636804c01778b3f82d6bd5d21972ad23fde7098c2aa1247d1" Oct 04 04:46:32 crc kubenswrapper[4770]: I1004 04:46:32.372413 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-9j55c" Oct 04 04:46:32 crc kubenswrapper[4770]: I1004 04:46:32.419747 4770 scope.go:117] "RemoveContainer" containerID="3bb4d9a5030e54ef869fc96c9b5ccaa9f4d533b795f36d996ee571280fa6ff27" Oct 04 04:46:32 crc kubenswrapper[4770]: I1004 04:46:32.426065 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-678599687f-9j55c"] Oct 04 04:46:32 crc kubenswrapper[4770]: I1004 04:46:32.433528 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-678599687f-9j55c"] Oct 04 04:46:33 crc kubenswrapper[4770]: I1004 04:46:33.687477 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" path="/var/lib/kubelet/pods/93e6caa9-b2cd-48e5-b332-5217fb8f110b/volumes" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.640617 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-678599687f-jn5sj"] Oct 04 04:46:35 crc kubenswrapper[4770]: E1004 04:46:35.642426 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" containerName="init" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.642510 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" containerName="init" Oct 04 04:46:35 crc kubenswrapper[4770]: E1004 04:46:35.642594 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" containerName="octavia-amphora-httpd" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.642667 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" containerName="octavia-amphora-httpd" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.642892 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e6caa9-b2cd-48e5-b332-5217fb8f110b" containerName="octavia-amphora-httpd" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.643936 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-jn5sj" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.650869 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.652859 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-jn5sj"] Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.674210 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:46:35 crc kubenswrapper[4770]: E1004 04:46:35.674527 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.817285 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/f4c708b4-49e9-4a43-9372-c932bb4bc473-amphora-image\") pod \"octavia-image-upload-678599687f-jn5sj\" (UID: \"f4c708b4-49e9-4a43-9372-c932bb4bc473\") " pod="openstack/octavia-image-upload-678599687f-jn5sj" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.817496 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f4c708b4-49e9-4a43-9372-c932bb4bc473-httpd-config\") pod \"octavia-image-upload-678599687f-jn5sj\" (UID: \"f4c708b4-49e9-4a43-9372-c932bb4bc473\") " pod="openstack/octavia-image-upload-678599687f-jn5sj" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.921082 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f4c708b4-49e9-4a43-9372-c932bb4bc473-httpd-config\") pod \"octavia-image-upload-678599687f-jn5sj\" (UID: \"f4c708b4-49e9-4a43-9372-c932bb4bc473\") " pod="openstack/octavia-image-upload-678599687f-jn5sj" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.921717 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/f4c708b4-49e9-4a43-9372-c932bb4bc473-amphora-image\") pod \"octavia-image-upload-678599687f-jn5sj\" (UID: \"f4c708b4-49e9-4a43-9372-c932bb4bc473\") " pod="openstack/octavia-image-upload-678599687f-jn5sj" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.922166 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/f4c708b4-49e9-4a43-9372-c932bb4bc473-amphora-image\") pod \"octavia-image-upload-678599687f-jn5sj\" (UID: \"f4c708b4-49e9-4a43-9372-c932bb4bc473\") " pod="openstack/octavia-image-upload-678599687f-jn5sj" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.936419 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f4c708b4-49e9-4a43-9372-c932bb4bc473-httpd-config\") pod \"octavia-image-upload-678599687f-jn5sj\" (UID: \"f4c708b4-49e9-4a43-9372-c932bb4bc473\") " pod="openstack/octavia-image-upload-678599687f-jn5sj" Oct 04 04:46:35 crc kubenswrapper[4770]: I1004 04:46:35.965775 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-jn5sj" Oct 04 04:46:36 crc kubenswrapper[4770]: I1004 04:46:36.442687 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-jn5sj"] Oct 04 04:46:37 crc kubenswrapper[4770]: I1004 04:46:37.422093 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-jn5sj" event={"ID":"f4c708b4-49e9-4a43-9372-c932bb4bc473","Type":"ContainerStarted","Data":"3a09efcf100df7884b7123ac0f07ab09f6260a0fcd77af18cea1b12b8f47f0d1"} Oct 04 04:46:38 crc kubenswrapper[4770]: I1004 04:46:38.437477 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-jn5sj" event={"ID":"f4c708b4-49e9-4a43-9372-c932bb4bc473","Type":"ContainerStarted","Data":"e83f99064a884391a06d879d430ad0340602e7494da4d4f68029169b2bc1b873"} Oct 04 04:46:39 crc kubenswrapper[4770]: I1004 04:46:39.448443 4770 generic.go:334] "Generic (PLEG): container finished" podID="f4c708b4-49e9-4a43-9372-c932bb4bc473" containerID="e83f99064a884391a06d879d430ad0340602e7494da4d4f68029169b2bc1b873" exitCode=0 Oct 04 04:46:39 crc kubenswrapper[4770]: I1004 04:46:39.448499 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-jn5sj" event={"ID":"f4c708b4-49e9-4a43-9372-c932bb4bc473","Type":"ContainerDied","Data":"e83f99064a884391a06d879d430ad0340602e7494da4d4f68029169b2bc1b873"} Oct 04 04:46:40 crc kubenswrapper[4770]: I1004 04:46:40.459452 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-jn5sj" event={"ID":"f4c708b4-49e9-4a43-9372-c932bb4bc473","Type":"ContainerStarted","Data":"876f9443c4ccc697c0c0ed252283a45825ff3bba443338120d7a42a428e31085"} Oct 04 04:46:40 crc kubenswrapper[4770]: I1004 04:46:40.478878 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-678599687f-jn5sj" podStartSLOduration=4.740719526 podStartE2EDuration="5.478853182s" podCreationTimestamp="2025-10-04 04:46:35 +0000 UTC" firstStartedPulling="2025-10-04 04:46:36.458390446 +0000 UTC m=+6207.750400158" lastFinishedPulling="2025-10-04 04:46:37.196524092 +0000 UTC m=+6208.488533814" observedRunningTime="2025-10-04 04:46:40.473668236 +0000 UTC m=+6211.765677948" watchObservedRunningTime="2025-10-04 04:46:40.478853182 +0000 UTC m=+6211.770862904" Oct 04 04:46:49 crc kubenswrapper[4770]: I1004 04:46:49.690092 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:46:49 crc kubenswrapper[4770]: E1004 04:46:49.690830 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:46:57 crc kubenswrapper[4770]: E1004 04:46:57.387274 4770 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.68:40266->38.102.83.68:36801: read tcp 38.102.83.68:40266->38.102.83.68:36801: read: connection reset by peer Oct 04 04:47:03 crc kubenswrapper[4770]: I1004 04:47:03.039202 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-jpf9d"] Oct 04 04:47:03 crc kubenswrapper[4770]: I1004 04:47:03.051092 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-jpf9d"] Oct 04 04:47:03 crc kubenswrapper[4770]: I1004 04:47:03.692771 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e9b805f-4925-4291-9a15-a7efac711ba5" path="/var/lib/kubelet/pods/2e9b805f-4925-4291-9a15-a7efac711ba5/volumes" Oct 04 04:47:04 crc kubenswrapper[4770]: I1004 04:47:04.027567 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-js2xf"] Oct 04 04:47:04 crc kubenswrapper[4770]: I1004 04:47:04.034807 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-js2xf"] Oct 04 04:47:04 crc kubenswrapper[4770]: I1004 04:47:04.044231 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-cwx5p"] Oct 04 04:47:04 crc kubenswrapper[4770]: I1004 04:47:04.052637 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-cwx5p"] Oct 04 04:47:04 crc kubenswrapper[4770]: I1004 04:47:04.674143 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:47:04 crc kubenswrapper[4770]: E1004 04:47:04.674543 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:47:05 crc kubenswrapper[4770]: I1004 04:47:05.685644 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c3bbbec-3d17-487c-b37d-4ed7254d14f5" path="/var/lib/kubelet/pods/2c3bbbec-3d17-487c-b37d-4ed7254d14f5/volumes" Oct 04 04:47:05 crc kubenswrapper[4770]: I1004 04:47:05.686568 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="322194fd-446b-49b0-9051-20edd7db6d8a" path="/var/lib/kubelet/pods/322194fd-446b-49b0-9051-20edd7db6d8a/volumes" Oct 04 04:47:14 crc kubenswrapper[4770]: I1004 04:47:14.039603 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-321b-account-create-q52jt"] Oct 04 04:47:14 crc kubenswrapper[4770]: I1004 04:47:14.050870 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-44d4-account-create-c8pxc"] Oct 04 04:47:14 crc kubenswrapper[4770]: I1004 04:47:14.061271 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-44d4-account-create-c8pxc"] Oct 04 04:47:14 crc kubenswrapper[4770]: I1004 04:47:14.071104 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-321b-account-create-q52jt"] Oct 04 04:47:15 crc kubenswrapper[4770]: I1004 04:47:15.032382 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-30a8-account-create-9sctc"] Oct 04 04:47:15 crc kubenswrapper[4770]: I1004 04:47:15.039651 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-30a8-account-create-9sctc"] Oct 04 04:47:15 crc kubenswrapper[4770]: I1004 04:47:15.690290 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49759e6a-459f-4355-ab2f-fb45039cc611" path="/var/lib/kubelet/pods/49759e6a-459f-4355-ab2f-fb45039cc611/volumes" Oct 04 04:47:15 crc kubenswrapper[4770]: I1004 04:47:15.690915 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66acc291-67a6-4e60-9c73-f9e7754853b7" path="/var/lib/kubelet/pods/66acc291-67a6-4e60-9c73-f9e7754853b7/volumes" Oct 04 04:47:15 crc kubenswrapper[4770]: I1004 04:47:15.691504 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2b431c5-ac55-489e-ab00-bb9ee7176b77" path="/var/lib/kubelet/pods/b2b431c5-ac55-489e-ab00-bb9ee7176b77/volumes" Oct 04 04:47:18 crc kubenswrapper[4770]: I1004 04:47:18.674755 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:47:18 crc kubenswrapper[4770]: E1004 04:47:18.675611 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.384869 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-55cf976f6f-xq5xj"] Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.386920 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.396456 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.396640 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.396806 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.396943 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-4p5r4" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.445875 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55cf976f6f-xq5xj"] Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.480374 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.480735 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="61d14385-5018-4129-9421-a92154491a56" containerName="glance-log" containerID="cri-o://9cbadeb8d48ac7bab37fe817e22037ac25e1ca21b04f53adb1faf1de0c47076a" gracePeriod=30 Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.481749 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="61d14385-5018-4129-9421-a92154491a56" containerName="glance-httpd" containerID="cri-o://4c463a78a6efad8325eb9f8ea8fcef2c9fb12b91d453d406495e3499bb0b2c38" gracePeriod=30 Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.506944 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-69cdc756fc-p6dcj"] Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.509156 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.527387 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69cdc756fc-p6dcj"] Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.534209 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bt65\" (UniqueName: \"kubernetes.io/projected/73a9a959-f316-4699-a1b6-9ecc07204706-kube-api-access-5bt65\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.534259 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-scripts\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.534338 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73a9a959-f316-4699-a1b6-9ecc07204706-logs\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.534387 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-config-data\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.534410 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73a9a959-f316-4699-a1b6-9ecc07204706-horizon-secret-key\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.555114 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.555522 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerName="glance-log" containerID="cri-o://b2dd5c11848132ac3c875a3c06f8e1e90acea5f608bfb558073bd62ce045db50" gracePeriod=30 Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.556081 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerName="glance-httpd" containerID="cri-o://5a5a976f6b1bcf77bbcbfdd404c05f4491480779ca55ecf4c5aa0037978849d7" gracePeriod=30 Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637096 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-config-data\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637372 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73a9a959-f316-4699-a1b6-9ecc07204706-horizon-secret-key\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637461 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bt65\" (UniqueName: \"kubernetes.io/projected/73a9a959-f316-4699-a1b6-9ecc07204706-kube-api-access-5bt65\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637496 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-config-data\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637536 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-scripts\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637569 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/193a3912-2337-4217-8da9-0b67f42a6220-horizon-secret-key\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637617 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/193a3912-2337-4217-8da9-0b67f42a6220-logs\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637662 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24pwr\" (UniqueName: \"kubernetes.io/projected/193a3912-2337-4217-8da9-0b67f42a6220-kube-api-access-24pwr\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637693 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-scripts\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.637714 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73a9a959-f316-4699-a1b6-9ecc07204706-logs\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.638208 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73a9a959-f316-4699-a1b6-9ecc07204706-logs\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.638254 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-config-data\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.639881 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-scripts\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.642692 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73a9a959-f316-4699-a1b6-9ecc07204706-horizon-secret-key\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.653247 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bt65\" (UniqueName: \"kubernetes.io/projected/73a9a959-f316-4699-a1b6-9ecc07204706-kube-api-access-5bt65\") pod \"horizon-55cf976f6f-xq5xj\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.732540 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.739513 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-config-data\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.739751 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/193a3912-2337-4217-8da9-0b67f42a6220-horizon-secret-key\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.739883 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/193a3912-2337-4217-8da9-0b67f42a6220-logs\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.740023 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24pwr\" (UniqueName: \"kubernetes.io/projected/193a3912-2337-4217-8da9-0b67f42a6220-kube-api-access-24pwr\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.740142 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-scripts\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.740770 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-scripts\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.740825 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-config-data\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.744178 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/193a3912-2337-4217-8da9-0b67f42a6220-horizon-secret-key\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.744500 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/193a3912-2337-4217-8da9-0b67f42a6220-logs\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.755365 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24pwr\" (UniqueName: \"kubernetes.io/projected/193a3912-2337-4217-8da9-0b67f42a6220-kube-api-access-24pwr\") pod \"horizon-69cdc756fc-p6dcj\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: E1004 04:47:26.801120 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61d14385_5018_4129_9421_a92154491a56.slice/crio-conmon-9cbadeb8d48ac7bab37fe817e22037ac25e1ca21b04f53adb1faf1de0c47076a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7783d14_4f52_47fa_8bb9_5a99cb5f7192.slice/crio-conmon-b2dd5c11848132ac3c875a3c06f8e1e90acea5f608bfb558073bd62ce045db50.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.854409 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.945228 4770 generic.go:334] "Generic (PLEG): container finished" podID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerID="b2dd5c11848132ac3c875a3c06f8e1e90acea5f608bfb558073bd62ce045db50" exitCode=143 Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.945291 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7783d14-4f52-47fa-8bb9-5a99cb5f7192","Type":"ContainerDied","Data":"b2dd5c11848132ac3c875a3c06f8e1e90acea5f608bfb558073bd62ce045db50"} Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.965919 4770 generic.go:334] "Generic (PLEG): container finished" podID="61d14385-5018-4129-9421-a92154491a56" containerID="9cbadeb8d48ac7bab37fe817e22037ac25e1ca21b04f53adb1faf1de0c47076a" exitCode=143 Oct 04 04:47:26 crc kubenswrapper[4770]: I1004 04:47:26.965974 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"61d14385-5018-4129-9421-a92154491a56","Type":"ContainerDied","Data":"9cbadeb8d48ac7bab37fe817e22037ac25e1ca21b04f53adb1faf1de0c47076a"} Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.071894 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55cf976f6f-xq5xj"] Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.113919 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-58946648d7-rhdw9"] Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.115702 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.148985 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58946648d7-rhdw9"] Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.265794 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbtzv\" (UniqueName: \"kubernetes.io/projected/e8d763f0-83cf-4dfd-975e-186f6149274e-kube-api-access-nbtzv\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.266199 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8d763f0-83cf-4dfd-975e-186f6149274e-logs\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.266237 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-scripts\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.266448 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-config-data\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.266512 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8d763f0-83cf-4dfd-975e-186f6149274e-horizon-secret-key\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.270291 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55cf976f6f-xq5xj"] Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.368210 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbtzv\" (UniqueName: \"kubernetes.io/projected/e8d763f0-83cf-4dfd-975e-186f6149274e-kube-api-access-nbtzv\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.368248 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8d763f0-83cf-4dfd-975e-186f6149274e-logs\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.368276 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-scripts\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.368341 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-config-data\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.368378 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8d763f0-83cf-4dfd-975e-186f6149274e-horizon-secret-key\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.369130 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-scripts\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.369403 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8d763f0-83cf-4dfd-975e-186f6149274e-logs\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.369724 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-config-data\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.373668 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8d763f0-83cf-4dfd-975e-186f6149274e-horizon-secret-key\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.386995 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbtzv\" (UniqueName: \"kubernetes.io/projected/e8d763f0-83cf-4dfd-975e-186f6149274e-kube-api-access-nbtzv\") pod \"horizon-58946648d7-rhdw9\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.428171 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69cdc756fc-p6dcj"] Oct 04 04:47:27 crc kubenswrapper[4770]: W1004 04:47:27.431670 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod193a3912_2337_4217_8da9_0b67f42a6220.slice/crio-e870184d89ee81ae63fb1c8be41eb91f8fdd5ce7d2bf11baec03b8d48b2e80a1 WatchSource:0}: Error finding container e870184d89ee81ae63fb1c8be41eb91f8fdd5ce7d2bf11baec03b8d48b2e80a1: Status 404 returned error can't find the container with id e870184d89ee81ae63fb1c8be41eb91f8fdd5ce7d2bf11baec03b8d48b2e80a1 Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.454846 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.935428 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58946648d7-rhdw9"] Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.935550 4770 scope.go:117] "RemoveContainer" containerID="9d38ba5616e3dc64bcc6918aa860e976782fff527fa61b4c5b028401c0739d8f" Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.980636 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58946648d7-rhdw9" event={"ID":"e8d763f0-83cf-4dfd-975e-186f6149274e","Type":"ContainerStarted","Data":"750ad3db6dbd6a2f90bfb7076f505d94aef03b74ab2f5e8050cd710ca31832a6"} Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.981997 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55cf976f6f-xq5xj" event={"ID":"73a9a959-f316-4699-a1b6-9ecc07204706","Type":"ContainerStarted","Data":"5a0702c1e6e776f8ba89dae733355db81afc95d4162ba2878e5778bcaf23ac9b"} Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.984212 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69cdc756fc-p6dcj" event={"ID":"193a3912-2337-4217-8da9-0b67f42a6220","Type":"ContainerStarted","Data":"e870184d89ee81ae63fb1c8be41eb91f8fdd5ce7d2bf11baec03b8d48b2e80a1"} Oct 04 04:47:27 crc kubenswrapper[4770]: I1004 04:47:27.995640 4770 scope.go:117] "RemoveContainer" containerID="ca76c67b2fc3d2abd623444b7f7d0deefd3a2f76d63a806f53f7693a0672e3b7" Oct 04 04:47:28 crc kubenswrapper[4770]: I1004 04:47:28.031145 4770 scope.go:117] "RemoveContainer" containerID="21a22851314d295d0c1bad2cc6a6b344fbf90d7e2a3f187db45faf02acadc4ab" Oct 04 04:47:28 crc kubenswrapper[4770]: I1004 04:47:28.081687 4770 scope.go:117] "RemoveContainer" containerID="1cb5eb88ff8f2484b1899265fafde1b2fce532ca1ea7bcba7c3cc23972f96d60" Oct 04 04:47:28 crc kubenswrapper[4770]: I1004 04:47:28.120498 4770 scope.go:117] "RemoveContainer" containerID="8c51acb01d6a9084308d50feea9fe0f3d2cafa976899da2f19aabac10056472a" Oct 04 04:47:28 crc kubenswrapper[4770]: I1004 04:47:28.147853 4770 scope.go:117] "RemoveContainer" containerID="a1ab32e4e547c7eed4f93a99d3690074701c89969f02b2c1201937c5a14cf3be" Oct 04 04:47:30 crc kubenswrapper[4770]: I1004 04:47:30.007643 4770 generic.go:334] "Generic (PLEG): container finished" podID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerID="5a5a976f6b1bcf77bbcbfdd404c05f4491480779ca55ecf4c5aa0037978849d7" exitCode=0 Oct 04 04:47:30 crc kubenswrapper[4770]: I1004 04:47:30.007714 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7783d14-4f52-47fa-8bb9-5a99cb5f7192","Type":"ContainerDied","Data":"5a5a976f6b1bcf77bbcbfdd404c05f4491480779ca55ecf4c5aa0037978849d7"} Oct 04 04:47:30 crc kubenswrapper[4770]: I1004 04:47:30.009743 4770 generic.go:334] "Generic (PLEG): container finished" podID="61d14385-5018-4129-9421-a92154491a56" containerID="4c463a78a6efad8325eb9f8ea8fcef2c9fb12b91d453d406495e3499bb0b2c38" exitCode=0 Oct 04 04:47:30 crc kubenswrapper[4770]: I1004 04:47:30.009771 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"61d14385-5018-4129-9421-a92154491a56","Type":"ContainerDied","Data":"4c463a78a6efad8325eb9f8ea8fcef2c9fb12b91d453d406495e3499bb0b2c38"} Oct 04 04:47:30 crc kubenswrapper[4770]: I1004 04:47:30.673805 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:47:30 crc kubenswrapper[4770]: E1004 04:47:30.674596 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:47:32 crc kubenswrapper[4770]: I1004 04:47:32.044317 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kd4rs"] Oct 04 04:47:32 crc kubenswrapper[4770]: I1004 04:47:32.055308 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kd4rs"] Oct 04 04:47:33 crc kubenswrapper[4770]: I1004 04:47:33.686502 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0e86553-7a36-44f8-b8fb-e21de6d2bda0" path="/var/lib/kubelet/pods/b0e86553-7a36-44f8-b8fb-e21de6d2bda0/volumes" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.063528 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58946648d7-rhdw9" event={"ID":"e8d763f0-83cf-4dfd-975e-186f6149274e","Type":"ContainerStarted","Data":"497f70a99994b7812505d59ac409e22473748a32eb26c4d48c5fc7a115fb75e1"} Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.065769 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69cdc756fc-p6dcj" event={"ID":"193a3912-2337-4217-8da9-0b67f42a6220","Type":"ContainerStarted","Data":"0b92cd5705bee2cb7dc44d7c45166bdb8831b73d134ca7f6985bb754df3f1f86"} Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.067149 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55cf976f6f-xq5xj" event={"ID":"73a9a959-f316-4699-a1b6-9ecc07204706","Type":"ContainerStarted","Data":"4b809e16e3be0c24e9a14252b0b1d76972a42248f01611e5e475be7e5b9a6426"} Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.311189 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.452486 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sppms\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-kube-api-access-sppms\") pod \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.452603 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-scripts\") pod \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.452658 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-config-data\") pod \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.453468 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-combined-ca-bundle\") pod \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.453532 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-logs\") pod \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.453584 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-httpd-run\") pod \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.453622 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-ceph\") pod \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\" (UID: \"f7783d14-4f52-47fa-8bb9-5a99cb5f7192\") " Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.454159 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f7783d14-4f52-47fa-8bb9-5a99cb5f7192" (UID: "f7783d14-4f52-47fa-8bb9-5a99cb5f7192"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.454547 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-logs" (OuterVolumeSpecName: "logs") pod "f7783d14-4f52-47fa-8bb9-5a99cb5f7192" (UID: "f7783d14-4f52-47fa-8bb9-5a99cb5f7192"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.459484 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-scripts" (OuterVolumeSpecName: "scripts") pod "f7783d14-4f52-47fa-8bb9-5a99cb5f7192" (UID: "f7783d14-4f52-47fa-8bb9-5a99cb5f7192"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.460130 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-kube-api-access-sppms" (OuterVolumeSpecName: "kube-api-access-sppms") pod "f7783d14-4f52-47fa-8bb9-5a99cb5f7192" (UID: "f7783d14-4f52-47fa-8bb9-5a99cb5f7192"). InnerVolumeSpecName "kube-api-access-sppms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.461280 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-ceph" (OuterVolumeSpecName: "ceph") pod "f7783d14-4f52-47fa-8bb9-5a99cb5f7192" (UID: "f7783d14-4f52-47fa-8bb9-5a99cb5f7192"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.480361 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7783d14-4f52-47fa-8bb9-5a99cb5f7192" (UID: "f7783d14-4f52-47fa-8bb9-5a99cb5f7192"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.519374 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-config-data" (OuterVolumeSpecName: "config-data") pod "f7783d14-4f52-47fa-8bb9-5a99cb5f7192" (UID: "f7783d14-4f52-47fa-8bb9-5a99cb5f7192"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.556568 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.556866 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.556878 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.556886 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.556897 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sppms\" (UniqueName: \"kubernetes.io/projected/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-kube-api-access-sppms\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.556908 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:35 crc kubenswrapper[4770]: I1004 04:47:35.556916 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7783d14-4f52-47fa-8bb9-5a99cb5f7192-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.085788 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58946648d7-rhdw9" event={"ID":"e8d763f0-83cf-4dfd-975e-186f6149274e","Type":"ContainerStarted","Data":"77b68824627eefb541868b796c23f6f96c21155c153751c88765a42c0fb63581"} Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.090299 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69cdc756fc-p6dcj" event={"ID":"193a3912-2337-4217-8da9-0b67f42a6220","Type":"ContainerStarted","Data":"5e322bda6df305b40256f3b062dc6710756a3c6fd9a6133f1a72bfd80be65ab3"} Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.095632 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7783d14-4f52-47fa-8bb9-5a99cb5f7192","Type":"ContainerDied","Data":"40d87c7a01f3bd728ef41dee6e00a6fa7d102d7be1e78050c4cbf3a1e7d0a378"} Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.095750 4770 scope.go:117] "RemoveContainer" containerID="5a5a976f6b1bcf77bbcbfdd404c05f4491480779ca55ecf4c5aa0037978849d7" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.095955 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.104494 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55cf976f6f-xq5xj" event={"ID":"73a9a959-f316-4699-a1b6-9ecc07204706","Type":"ContainerStarted","Data":"9e2e626e8754ff1b366c740a7edb6bbfc43af5a1d85795c2d43c084550a9e2c1"} Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.104845 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-55cf976f6f-xq5xj" podUID="73a9a959-f316-4699-a1b6-9ecc07204706" containerName="horizon-log" containerID="cri-o://4b809e16e3be0c24e9a14252b0b1d76972a42248f01611e5e475be7e5b9a6426" gracePeriod=30 Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.104907 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-55cf976f6f-xq5xj" podUID="73a9a959-f316-4699-a1b6-9ecc07204706" containerName="horizon" containerID="cri-o://9e2e626e8754ff1b366c740a7edb6bbfc43af5a1d85795c2d43c084550a9e2c1" gracePeriod=30 Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.128129 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-58946648d7-rhdw9" podStartSLOduration=2.381742635 podStartE2EDuration="9.128103847s" podCreationTimestamp="2025-10-04 04:47:27 +0000 UTC" firstStartedPulling="2025-10-04 04:47:27.940327663 +0000 UTC m=+6259.232337385" lastFinishedPulling="2025-10-04 04:47:34.686688885 +0000 UTC m=+6265.978698597" observedRunningTime="2025-10-04 04:47:36.108280188 +0000 UTC m=+6267.400289910" watchObservedRunningTime="2025-10-04 04:47:36.128103847 +0000 UTC m=+6267.420113559" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.147203 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.155931 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.168910 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-69cdc756fc-p6dcj" podStartSLOduration=2.8671365829999997 podStartE2EDuration="10.168894053s" podCreationTimestamp="2025-10-04 04:47:26 +0000 UTC" firstStartedPulling="2025-10-04 04:47:27.433738042 +0000 UTC m=+6258.725747754" lastFinishedPulling="2025-10-04 04:47:34.735495512 +0000 UTC m=+6266.027505224" observedRunningTime="2025-10-04 04:47:36.16033795 +0000 UTC m=+6267.452347662" watchObservedRunningTime="2025-10-04 04:47:36.168894053 +0000 UTC m=+6267.460903765" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.170419 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:47:36 crc kubenswrapper[4770]: E1004 04:47:36.170829 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerName="glance-httpd" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.170847 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerName="glance-httpd" Oct 04 04:47:36 crc kubenswrapper[4770]: E1004 04:47:36.170875 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerName="glance-log" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.170881 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerName="glance-log" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.171563 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerName="glance-httpd" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.171586 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" containerName="glance-log" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.172742 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.173366 4770 scope.go:117] "RemoveContainer" containerID="b2dd5c11848132ac3c875a3c06f8e1e90acea5f608bfb558073bd62ce045db50" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.175766 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.246596 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.249597 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-55cf976f6f-xq5xj" podStartSLOduration=2.83030778 podStartE2EDuration="10.249578683s" podCreationTimestamp="2025-10-04 04:47:26 +0000 UTC" firstStartedPulling="2025-10-04 04:47:27.267104544 +0000 UTC m=+6258.559114256" lastFinishedPulling="2025-10-04 04:47:34.686375407 +0000 UTC m=+6265.978385159" observedRunningTime="2025-10-04 04:47:36.1909612 +0000 UTC m=+6267.482970912" watchObservedRunningTime="2025-10-04 04:47:36.249578683 +0000 UTC m=+6267.541588395" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.268847 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdvpr\" (UniqueName: \"kubernetes.io/projected/be23d13e-755b-4ad3-883c-e52f4bb1146c-kube-api-access-sdvpr\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.268937 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be23d13e-755b-4ad3-883c-e52f4bb1146c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.268972 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be23d13e-755b-4ad3-883c-e52f4bb1146c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.269017 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be23d13e-755b-4ad3-883c-e52f4bb1146c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.269050 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be23d13e-755b-4ad3-883c-e52f4bb1146c-logs\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.269090 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/be23d13e-755b-4ad3-883c-e52f4bb1146c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.269111 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be23d13e-755b-4ad3-883c-e52f4bb1146c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.318287 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.370363 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be23d13e-755b-4ad3-883c-e52f4bb1146c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.370485 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be23d13e-755b-4ad3-883c-e52f4bb1146c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.370529 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be23d13e-755b-4ad3-883c-e52f4bb1146c-logs\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.370597 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/be23d13e-755b-4ad3-883c-e52f4bb1146c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.370643 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be23d13e-755b-4ad3-883c-e52f4bb1146c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.370748 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdvpr\" (UniqueName: \"kubernetes.io/projected/be23d13e-755b-4ad3-883c-e52f4bb1146c-kube-api-access-sdvpr\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.370818 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be23d13e-755b-4ad3-883c-e52f4bb1146c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.372051 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be23d13e-755b-4ad3-883c-e52f4bb1146c-logs\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.372124 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be23d13e-755b-4ad3-883c-e52f4bb1146c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.378424 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/be23d13e-755b-4ad3-883c-e52f4bb1146c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.379360 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be23d13e-755b-4ad3-883c-e52f4bb1146c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.380166 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be23d13e-755b-4ad3-883c-e52f4bb1146c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.396609 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be23d13e-755b-4ad3-883c-e52f4bb1146c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.398639 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdvpr\" (UniqueName: \"kubernetes.io/projected/be23d13e-755b-4ad3-883c-e52f4bb1146c-kube-api-access-sdvpr\") pod \"glance-default-internal-api-0\" (UID: \"be23d13e-755b-4ad3-883c-e52f4bb1146c\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.477822 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-scripts\") pod \"61d14385-5018-4129-9421-a92154491a56\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.477960 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-ceph\") pod \"61d14385-5018-4129-9421-a92154491a56\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.478069 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fx8qt\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-kube-api-access-fx8qt\") pod \"61d14385-5018-4129-9421-a92154491a56\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.478153 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-httpd-run\") pod \"61d14385-5018-4129-9421-a92154491a56\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.478189 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-config-data\") pod \"61d14385-5018-4129-9421-a92154491a56\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.478207 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-combined-ca-bundle\") pod \"61d14385-5018-4129-9421-a92154491a56\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.478229 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-logs\") pod \"61d14385-5018-4129-9421-a92154491a56\" (UID: \"61d14385-5018-4129-9421-a92154491a56\") " Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.480174 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "61d14385-5018-4129-9421-a92154491a56" (UID: "61d14385-5018-4129-9421-a92154491a56"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.480192 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-logs" (OuterVolumeSpecName: "logs") pod "61d14385-5018-4129-9421-a92154491a56" (UID: "61d14385-5018-4129-9421-a92154491a56"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.481450 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-scripts" (OuterVolumeSpecName: "scripts") pod "61d14385-5018-4129-9421-a92154491a56" (UID: "61d14385-5018-4129-9421-a92154491a56"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.483985 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-kube-api-access-fx8qt" (OuterVolumeSpecName: "kube-api-access-fx8qt") pod "61d14385-5018-4129-9421-a92154491a56" (UID: "61d14385-5018-4129-9421-a92154491a56"). InnerVolumeSpecName "kube-api-access-fx8qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.487226 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-ceph" (OuterVolumeSpecName: "ceph") pod "61d14385-5018-4129-9421-a92154491a56" (UID: "61d14385-5018-4129-9421-a92154491a56"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.515236 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61d14385-5018-4129-9421-a92154491a56" (UID: "61d14385-5018-4129-9421-a92154491a56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.542452 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-config-data" (OuterVolumeSpecName: "config-data") pod "61d14385-5018-4129-9421-a92154491a56" (UID: "61d14385-5018-4129-9421-a92154491a56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.597534 4770 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.597794 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.597807 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.597821 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61d14385-5018-4129-9421-a92154491a56-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.597832 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61d14385-5018-4129-9421-a92154491a56-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.597843 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.597854 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fx8qt\" (UniqueName: \"kubernetes.io/projected/61d14385-5018-4129-9421-a92154491a56-kube-api-access-fx8qt\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.601490 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.733473 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.855260 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:36 crc kubenswrapper[4770]: I1004 04:47:36.855592 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.138069 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"61d14385-5018-4129-9421-a92154491a56","Type":"ContainerDied","Data":"a0ff65de50f98b756aac9bba08dc7a80022391746c229dfaa27788007baa6c27"} Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.138124 4770 scope.go:117] "RemoveContainer" containerID="4c463a78a6efad8325eb9f8ea8fcef2c9fb12b91d453d406495e3499bb0b2c38" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.138295 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.212243 4770 scope.go:117] "RemoveContainer" containerID="9cbadeb8d48ac7bab37fe817e22037ac25e1ca21b04f53adb1faf1de0c47076a" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.216084 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.234426 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.257974 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:47:37 crc kubenswrapper[4770]: E1004 04:47:37.258474 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d14385-5018-4129-9421-a92154491a56" containerName="glance-log" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.258499 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d14385-5018-4129-9421-a92154491a56" containerName="glance-log" Oct 04 04:47:37 crc kubenswrapper[4770]: E1004 04:47:37.258547 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d14385-5018-4129-9421-a92154491a56" containerName="glance-httpd" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.258556 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d14385-5018-4129-9421-a92154491a56" containerName="glance-httpd" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.258790 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d14385-5018-4129-9421-a92154491a56" containerName="glance-log" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.258830 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d14385-5018-4129-9421-a92154491a56" containerName="glance-httpd" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.260326 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.265475 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.275991 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.381668 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.422828 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-config-data\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.423057 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-scripts\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.423109 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zdkr\" (UniqueName: \"kubernetes.io/projected/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-kube-api-access-5zdkr\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.423159 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.423400 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-ceph\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.423506 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-logs\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.423730 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.455764 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.455817 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.525871 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-config-data\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.525975 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-scripts\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.525998 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zdkr\" (UniqueName: \"kubernetes.io/projected/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-kube-api-access-5zdkr\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.526041 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.526076 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-ceph\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.526098 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-logs\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.526156 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.526602 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.529366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-logs\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.532757 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.533059 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-config-data\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.534174 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-ceph\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.534891 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-scripts\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.551214 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zdkr\" (UniqueName: \"kubernetes.io/projected/6b5e1455-d5f5-49d4-84a5-2aadce69bc3e-kube-api-access-5zdkr\") pod \"glance-default-external-api-0\" (UID: \"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e\") " pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.602670 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.684938 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d14385-5018-4129-9421-a92154491a56" path="/var/lib/kubelet/pods/61d14385-5018-4129-9421-a92154491a56/volumes" Oct 04 04:47:37 crc kubenswrapper[4770]: I1004 04:47:37.686402 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7783d14-4f52-47fa-8bb9-5a99cb5f7192" path="/var/lib/kubelet/pods/f7783d14-4f52-47fa-8bb9-5a99cb5f7192/volumes" Oct 04 04:47:38 crc kubenswrapper[4770]: I1004 04:47:38.160424 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be23d13e-755b-4ad3-883c-e52f4bb1146c","Type":"ContainerStarted","Data":"d74ff3487073161dd8ddb74a02c2573bbe64357949133590a0e169eae93c4988"} Oct 04 04:47:38 crc kubenswrapper[4770]: I1004 04:47:38.184798 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:47:38 crc kubenswrapper[4770]: W1004 04:47:38.193283 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b5e1455_d5f5_49d4_84a5_2aadce69bc3e.slice/crio-e220b8cc700b67e16cb58f3f732aaf94389ee48865f604851de62b423d6c6093 WatchSource:0}: Error finding container e220b8cc700b67e16cb58f3f732aaf94389ee48865f604851de62b423d6c6093: Status 404 returned error can't find the container with id e220b8cc700b67e16cb58f3f732aaf94389ee48865f604851de62b423d6c6093 Oct 04 04:47:39 crc kubenswrapper[4770]: I1004 04:47:39.193716 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be23d13e-755b-4ad3-883c-e52f4bb1146c","Type":"ContainerStarted","Data":"a7a0967a1d9d30a348b854d15e1189fd8c9746e474f426e183468a39e96227e2"} Oct 04 04:47:39 crc kubenswrapper[4770]: I1004 04:47:39.203848 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e","Type":"ContainerStarted","Data":"b997396c5d95d5d92539f581afaa78e79733acb74bab1ed60c58c0ad18b7b9fd"} Oct 04 04:47:39 crc kubenswrapper[4770]: I1004 04:47:39.203958 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e","Type":"ContainerStarted","Data":"e220b8cc700b67e16cb58f3f732aaf94389ee48865f604851de62b423d6c6093"} Oct 04 04:47:40 crc kubenswrapper[4770]: I1004 04:47:40.222177 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b5e1455-d5f5-49d4-84a5-2aadce69bc3e","Type":"ContainerStarted","Data":"5af0c60510fe975c2e55c9ea93e8506e827548ffa1de09c8629741cc7b53a775"} Oct 04 04:47:40 crc kubenswrapper[4770]: I1004 04:47:40.236169 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be23d13e-755b-4ad3-883c-e52f4bb1146c","Type":"ContainerStarted","Data":"8914c1e906f1d0d898b9f42ae30196fe7170cf6ce2675f98c1d34c5cf16570d0"} Oct 04 04:47:40 crc kubenswrapper[4770]: I1004 04:47:40.258534 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.258505689 podStartE2EDuration="3.258505689s" podCreationTimestamp="2025-10-04 04:47:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:47:40.246185677 +0000 UTC m=+6271.538195409" watchObservedRunningTime="2025-10-04 04:47:40.258505689 +0000 UTC m=+6271.550515411" Oct 04 04:47:40 crc kubenswrapper[4770]: I1004 04:47:40.286710 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.286684726 podStartE2EDuration="4.286684726s" podCreationTimestamp="2025-10-04 04:47:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:47:40.269241769 +0000 UTC m=+6271.561251491" watchObservedRunningTime="2025-10-04 04:47:40.286684726 +0000 UTC m=+6271.578694448" Oct 04 04:47:44 crc kubenswrapper[4770]: I1004 04:47:44.673898 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:47:44 crc kubenswrapper[4770]: E1004 04:47:44.674630 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:47:46 crc kubenswrapper[4770]: I1004 04:47:46.619888 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:46 crc kubenswrapper[4770]: I1004 04:47:46.620197 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:46 crc kubenswrapper[4770]: I1004 04:47:46.664980 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:46 crc kubenswrapper[4770]: I1004 04:47:46.671877 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:46 crc kubenswrapper[4770]: I1004 04:47:46.856985 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-69cdc756fc-p6dcj" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Oct 04 04:47:47 crc kubenswrapper[4770]: I1004 04:47:47.316491 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:47 crc kubenswrapper[4770]: I1004 04:47:47.316529 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:47 crc kubenswrapper[4770]: I1004 04:47:47.457569 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-58946648d7-rhdw9" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 04 04:47:47 crc kubenswrapper[4770]: I1004 04:47:47.603621 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:47:47 crc kubenswrapper[4770]: I1004 04:47:47.603696 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:47:47 crc kubenswrapper[4770]: I1004 04:47:47.638632 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:47:47 crc kubenswrapper[4770]: I1004 04:47:47.670584 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:47:48 crc kubenswrapper[4770]: I1004 04:47:48.324744 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:47:48 crc kubenswrapper[4770]: I1004 04:47:48.324804 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:47:49 crc kubenswrapper[4770]: I1004 04:47:49.556734 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:49 crc kubenswrapper[4770]: I1004 04:47:49.557149 4770 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:47:49 crc kubenswrapper[4770]: I1004 04:47:49.581147 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.086367 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-68tgg"] Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.088912 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.095856 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68tgg"] Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.214684 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjm45\" (UniqueName: \"kubernetes.io/projected/95b3f1d6-060c-4d9d-931f-f08f273d556e-kube-api-access-tjm45\") pod \"redhat-marketplace-68tgg\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.215085 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-catalog-content\") pod \"redhat-marketplace-68tgg\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.215134 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-utilities\") pod \"redhat-marketplace-68tgg\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.317544 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjm45\" (UniqueName: \"kubernetes.io/projected/95b3f1d6-060c-4d9d-931f-f08f273d556e-kube-api-access-tjm45\") pod \"redhat-marketplace-68tgg\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.317607 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-catalog-content\") pod \"redhat-marketplace-68tgg\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.317657 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-utilities\") pod \"redhat-marketplace-68tgg\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.318222 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-utilities\") pod \"redhat-marketplace-68tgg\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.318317 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-catalog-content\") pod \"redhat-marketplace-68tgg\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.350638 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjm45\" (UniqueName: \"kubernetes.io/projected/95b3f1d6-060c-4d9d-931f-f08f273d556e-kube-api-access-tjm45\") pod \"redhat-marketplace-68tgg\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.420423 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.830904 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.831210 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:47:50 crc kubenswrapper[4770]: I1004 04:47:50.940508 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-68tgg"] Oct 04 04:47:51 crc kubenswrapper[4770]: I1004 04:47:51.044576 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdzqf"] Oct 04 04:47:51 crc kubenswrapper[4770]: I1004 04:47:51.054470 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xdzqf"] Oct 04 04:47:51 crc kubenswrapper[4770]: I1004 04:47:51.389710 4770 generic.go:334] "Generic (PLEG): container finished" podID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerID="b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3" exitCode=0 Oct 04 04:47:51 crc kubenswrapper[4770]: I1004 04:47:51.389801 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68tgg" event={"ID":"95b3f1d6-060c-4d9d-931f-f08f273d556e","Type":"ContainerDied","Data":"b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3"} Oct 04 04:47:51 crc kubenswrapper[4770]: I1004 04:47:51.389833 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68tgg" event={"ID":"95b3f1d6-060c-4d9d-931f-f08f273d556e","Type":"ContainerStarted","Data":"44d52bd4b8c2b59787a8222f4a1eaba460971d57f57bc6529ca2a08bd72b58f8"} Oct 04 04:47:51 crc kubenswrapper[4770]: I1004 04:47:51.404696 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:47:51 crc kubenswrapper[4770]: I1004 04:47:51.693543 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c093be9-71c5-4478-84b2-da4d87396fc9" path="/var/lib/kubelet/pods/3c093be9-71c5-4478-84b2-da4d87396fc9/volumes" Oct 04 04:47:52 crc kubenswrapper[4770]: I1004 04:47:52.025865 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-gpw5q"] Oct 04 04:47:52 crc kubenswrapper[4770]: I1004 04:47:52.037173 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-gpw5q"] Oct 04 04:47:53 crc kubenswrapper[4770]: I1004 04:47:53.410261 4770 generic.go:334] "Generic (PLEG): container finished" podID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerID="8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea" exitCode=0 Oct 04 04:47:53 crc kubenswrapper[4770]: I1004 04:47:53.410317 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68tgg" event={"ID":"95b3f1d6-060c-4d9d-931f-f08f273d556e","Type":"ContainerDied","Data":"8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea"} Oct 04 04:47:53 crc kubenswrapper[4770]: I1004 04:47:53.685462 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38e50e99-f55d-47a6-ab20-fc4c37c91852" path="/var/lib/kubelet/pods/38e50e99-f55d-47a6-ab20-fc4c37c91852/volumes" Oct 04 04:47:54 crc kubenswrapper[4770]: I1004 04:47:54.421861 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68tgg" event={"ID":"95b3f1d6-060c-4d9d-931f-f08f273d556e","Type":"ContainerStarted","Data":"474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3"} Oct 04 04:47:54 crc kubenswrapper[4770]: I1004 04:47:54.448992 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-68tgg" podStartSLOduration=1.669095067 podStartE2EDuration="4.448975535s" podCreationTimestamp="2025-10-04 04:47:50 +0000 UTC" firstStartedPulling="2025-10-04 04:47:51.404364032 +0000 UTC m=+6282.696373744" lastFinishedPulling="2025-10-04 04:47:54.18424446 +0000 UTC m=+6285.476254212" observedRunningTime="2025-10-04 04:47:54.443463641 +0000 UTC m=+6285.735473353" watchObservedRunningTime="2025-10-04 04:47:54.448975535 +0000 UTC m=+6285.740985247" Oct 04 04:47:55 crc kubenswrapper[4770]: I1004 04:47:55.673946 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:47:55 crc kubenswrapper[4770]: E1004 04:47:55.674658 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:47:58 crc kubenswrapper[4770]: I1004 04:47:58.932731 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:47:59 crc kubenswrapper[4770]: I1004 04:47:59.452963 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:48:00 crc kubenswrapper[4770]: I1004 04:48:00.420806 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:48:00 crc kubenswrapper[4770]: I1004 04:48:00.421180 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:48:00 crc kubenswrapper[4770]: I1004 04:48:00.493178 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:48:00 crc kubenswrapper[4770]: I1004 04:48:00.548269 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:48:00 crc kubenswrapper[4770]: I1004 04:48:00.671081 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:48:00 crc kubenswrapper[4770]: I1004 04:48:00.744781 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68tgg"] Oct 04 04:48:01 crc kubenswrapper[4770]: I1004 04:48:01.016968 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:48:01 crc kubenswrapper[4770]: I1004 04:48:01.072261 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69cdc756fc-p6dcj"] Oct 04 04:48:01 crc kubenswrapper[4770]: I1004 04:48:01.499056 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69cdc756fc-p6dcj" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon-log" containerID="cri-o://0b92cd5705bee2cb7dc44d7c45166bdb8831b73d134ca7f6985bb754df3f1f86" gracePeriod=30 Oct 04 04:48:01 crc kubenswrapper[4770]: I1004 04:48:01.499149 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69cdc756fc-p6dcj" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon" containerID="cri-o://5e322bda6df305b40256f3b062dc6710756a3c6fd9a6133f1a72bfd80be65ab3" gracePeriod=30 Oct 04 04:48:02 crc kubenswrapper[4770]: I1004 04:48:02.505271 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-68tgg" podUID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerName="registry-server" containerID="cri-o://474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3" gracePeriod=2 Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.113268 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.211618 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-catalog-content\") pod \"95b3f1d6-060c-4d9d-931f-f08f273d556e\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.211948 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjm45\" (UniqueName: \"kubernetes.io/projected/95b3f1d6-060c-4d9d-931f-f08f273d556e-kube-api-access-tjm45\") pod \"95b3f1d6-060c-4d9d-931f-f08f273d556e\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.212212 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-utilities\") pod \"95b3f1d6-060c-4d9d-931f-f08f273d556e\" (UID: \"95b3f1d6-060c-4d9d-931f-f08f273d556e\") " Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.213170 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-utilities" (OuterVolumeSpecName: "utilities") pod "95b3f1d6-060c-4d9d-931f-f08f273d556e" (UID: "95b3f1d6-060c-4d9d-931f-f08f273d556e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.220398 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95b3f1d6-060c-4d9d-931f-f08f273d556e-kube-api-access-tjm45" (OuterVolumeSpecName: "kube-api-access-tjm45") pod "95b3f1d6-060c-4d9d-931f-f08f273d556e" (UID: "95b3f1d6-060c-4d9d-931f-f08f273d556e"). InnerVolumeSpecName "kube-api-access-tjm45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.225799 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95b3f1d6-060c-4d9d-931f-f08f273d556e" (UID: "95b3f1d6-060c-4d9d-931f-f08f273d556e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.315754 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjm45\" (UniqueName: \"kubernetes.io/projected/95b3f1d6-060c-4d9d-931f-f08f273d556e-kube-api-access-tjm45\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.315850 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.315893 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b3f1d6-060c-4d9d-931f-f08f273d556e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.520305 4770 generic.go:334] "Generic (PLEG): container finished" podID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerID="474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3" exitCode=0 Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.520358 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68tgg" event={"ID":"95b3f1d6-060c-4d9d-931f-f08f273d556e","Type":"ContainerDied","Data":"474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3"} Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.520395 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-68tgg" event={"ID":"95b3f1d6-060c-4d9d-931f-f08f273d556e","Type":"ContainerDied","Data":"44d52bd4b8c2b59787a8222f4a1eaba460971d57f57bc6529ca2a08bd72b58f8"} Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.520420 4770 scope.go:117] "RemoveContainer" containerID="474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.520454 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-68tgg" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.551376 4770 scope.go:117] "RemoveContainer" containerID="8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.568530 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-68tgg"] Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.576663 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-68tgg"] Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.589561 4770 scope.go:117] "RemoveContainer" containerID="b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.629094 4770 scope.go:117] "RemoveContainer" containerID="474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3" Oct 04 04:48:03 crc kubenswrapper[4770]: E1004 04:48:03.629644 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3\": container with ID starting with 474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3 not found: ID does not exist" containerID="474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.629701 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3"} err="failed to get container status \"474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3\": rpc error: code = NotFound desc = could not find container \"474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3\": container with ID starting with 474e18a0bdfbab6d8447b0b5888aff50383a01a22997c55e6dee8d955771e1c3 not found: ID does not exist" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.629728 4770 scope.go:117] "RemoveContainer" containerID="8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea" Oct 04 04:48:03 crc kubenswrapper[4770]: E1004 04:48:03.629959 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea\": container with ID starting with 8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea not found: ID does not exist" containerID="8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.629994 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea"} err="failed to get container status \"8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea\": rpc error: code = NotFound desc = could not find container \"8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea\": container with ID starting with 8d448af217ab84e01aa1476238a23a9955e0d892663809522381246c797193ea not found: ID does not exist" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.630034 4770 scope.go:117] "RemoveContainer" containerID="b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3" Oct 04 04:48:03 crc kubenswrapper[4770]: E1004 04:48:03.630290 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3\": container with ID starting with b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3 not found: ID does not exist" containerID="b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.630314 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3"} err="failed to get container status \"b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3\": rpc error: code = NotFound desc = could not find container \"b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3\": container with ID starting with b9e47af47157353397c467fc2ead308d9bd9198d4a58c6d92f0a44609cb4ced3 not found: ID does not exist" Oct 04 04:48:03 crc kubenswrapper[4770]: I1004 04:48:03.690298 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95b3f1d6-060c-4d9d-931f-f08f273d556e" path="/var/lib/kubelet/pods/95b3f1d6-060c-4d9d-931f-f08f273d556e/volumes" Oct 04 04:48:05 crc kubenswrapper[4770]: I1004 04:48:05.544483 4770 generic.go:334] "Generic (PLEG): container finished" podID="193a3912-2337-4217-8da9-0b67f42a6220" containerID="5e322bda6df305b40256f3b062dc6710756a3c6fd9a6133f1a72bfd80be65ab3" exitCode=0 Oct 04 04:48:05 crc kubenswrapper[4770]: I1004 04:48:05.544572 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69cdc756fc-p6dcj" event={"ID":"193a3912-2337-4217-8da9-0b67f42a6220","Type":"ContainerDied","Data":"5e322bda6df305b40256f3b062dc6710756a3c6fd9a6133f1a72bfd80be65ab3"} Oct 04 04:48:06 crc kubenswrapper[4770]: I1004 04:48:06.558762 4770 generic.go:334] "Generic (PLEG): container finished" podID="73a9a959-f316-4699-a1b6-9ecc07204706" containerID="9e2e626e8754ff1b366c740a7edb6bbfc43af5a1d85795c2d43c084550a9e2c1" exitCode=137 Oct 04 04:48:06 crc kubenswrapper[4770]: I1004 04:48:06.558795 4770 generic.go:334] "Generic (PLEG): container finished" podID="73a9a959-f316-4699-a1b6-9ecc07204706" containerID="4b809e16e3be0c24e9a14252b0b1d76972a42248f01611e5e475be7e5b9a6426" exitCode=137 Oct 04 04:48:06 crc kubenswrapper[4770]: I1004 04:48:06.558815 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55cf976f6f-xq5xj" event={"ID":"73a9a959-f316-4699-a1b6-9ecc07204706","Type":"ContainerDied","Data":"9e2e626e8754ff1b366c740a7edb6bbfc43af5a1d85795c2d43c084550a9e2c1"} Oct 04 04:48:06 crc kubenswrapper[4770]: I1004 04:48:06.558840 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55cf976f6f-xq5xj" event={"ID":"73a9a959-f316-4699-a1b6-9ecc07204706","Type":"ContainerDied","Data":"4b809e16e3be0c24e9a14252b0b1d76972a42248f01611e5e475be7e5b9a6426"} Oct 04 04:48:06 crc kubenswrapper[4770]: I1004 04:48:06.674321 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:48:06 crc kubenswrapper[4770]: E1004 04:48:06.674741 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:06.856870 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-69cdc756fc-p6dcj" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.018071 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.195789 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-scripts\") pod \"73a9a959-f316-4699-a1b6-9ecc07204706\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.195904 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73a9a959-f316-4699-a1b6-9ecc07204706-logs\") pod \"73a9a959-f316-4699-a1b6-9ecc07204706\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.196030 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-config-data\") pod \"73a9a959-f316-4699-a1b6-9ecc07204706\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.196222 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73a9a959-f316-4699-a1b6-9ecc07204706-horizon-secret-key\") pod \"73a9a959-f316-4699-a1b6-9ecc07204706\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.196246 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bt65\" (UniqueName: \"kubernetes.io/projected/73a9a959-f316-4699-a1b6-9ecc07204706-kube-api-access-5bt65\") pod \"73a9a959-f316-4699-a1b6-9ecc07204706\" (UID: \"73a9a959-f316-4699-a1b6-9ecc07204706\") " Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.196602 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73a9a959-f316-4699-a1b6-9ecc07204706-logs" (OuterVolumeSpecName: "logs") pod "73a9a959-f316-4699-a1b6-9ecc07204706" (UID: "73a9a959-f316-4699-a1b6-9ecc07204706"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.196987 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73a9a959-f316-4699-a1b6-9ecc07204706-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.202223 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73a9a959-f316-4699-a1b6-9ecc07204706-kube-api-access-5bt65" (OuterVolumeSpecName: "kube-api-access-5bt65") pod "73a9a959-f316-4699-a1b6-9ecc07204706" (UID: "73a9a959-f316-4699-a1b6-9ecc07204706"). InnerVolumeSpecName "kube-api-access-5bt65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.223791 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-scripts" (OuterVolumeSpecName: "scripts") pod "73a9a959-f316-4699-a1b6-9ecc07204706" (UID: "73a9a959-f316-4699-a1b6-9ecc07204706"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.225629 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73a9a959-f316-4699-a1b6-9ecc07204706-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "73a9a959-f316-4699-a1b6-9ecc07204706" (UID: "73a9a959-f316-4699-a1b6-9ecc07204706"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.225813 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-config-data" (OuterVolumeSpecName: "config-data") pod "73a9a959-f316-4699-a1b6-9ecc07204706" (UID: "73a9a959-f316-4699-a1b6-9ecc07204706"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.298814 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.298859 4770 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73a9a959-f316-4699-a1b6-9ecc07204706-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.298882 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bt65\" (UniqueName: \"kubernetes.io/projected/73a9a959-f316-4699-a1b6-9ecc07204706-kube-api-access-5bt65\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.298900 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73a9a959-f316-4699-a1b6-9ecc07204706-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.570335 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55cf976f6f-xq5xj" event={"ID":"73a9a959-f316-4699-a1b6-9ecc07204706","Type":"ContainerDied","Data":"5a0702c1e6e776f8ba89dae733355db81afc95d4162ba2878e5778bcaf23ac9b"} Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.570692 4770 scope.go:117] "RemoveContainer" containerID="9e2e626e8754ff1b366c740a7edb6bbfc43af5a1d85795c2d43c084550a9e2c1" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.570385 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55cf976f6f-xq5xj" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.609167 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55cf976f6f-xq5xj"] Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.618600 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-55cf976f6f-xq5xj"] Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.688137 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73a9a959-f316-4699-a1b6-9ecc07204706" path="/var/lib/kubelet/pods/73a9a959-f316-4699-a1b6-9ecc07204706/volumes" Oct 04 04:48:08 crc kubenswrapper[4770]: I1004 04:48:07.733111 4770 scope.go:117] "RemoveContainer" containerID="4b809e16e3be0c24e9a14252b0b1d76972a42248f01611e5e475be7e5b9a6426" Oct 04 04:48:12 crc kubenswrapper[4770]: I1004 04:48:12.054420 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-c6n99"] Oct 04 04:48:12 crc kubenswrapper[4770]: I1004 04:48:12.064310 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-c6n99"] Oct 04 04:48:13 crc kubenswrapper[4770]: I1004 04:48:13.687175 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34e7fc27-dfe7-4bb9-b6fc-b2fabc466819" path="/var/lib/kubelet/pods/34e7fc27-dfe7-4bb9-b6fc-b2fabc466819/volumes" Oct 04 04:48:16 crc kubenswrapper[4770]: I1004 04:48:16.856441 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-69cdc756fc-p6dcj" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Oct 04 04:48:19 crc kubenswrapper[4770]: I1004 04:48:19.680557 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:48:19 crc kubenswrapper[4770]: E1004 04:48:19.681471 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:48:26 crc kubenswrapper[4770]: I1004 04:48:26.859686 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-69cdc756fc-p6dcj" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Oct 04 04:48:26 crc kubenswrapper[4770]: I1004 04:48:26.860587 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:48:28 crc kubenswrapper[4770]: I1004 04:48:28.352436 4770 scope.go:117] "RemoveContainer" containerID="abb7af7395e58bb4795f53134cc7c15f7698cbe539b0c93e346617e09db55e5c" Oct 04 04:48:28 crc kubenswrapper[4770]: I1004 04:48:28.410591 4770 scope.go:117] "RemoveContainer" containerID="7ecc1d7c90290f90a4ef3e75d7a375fce97a408ebc53a7ea3a700a530d8d1ff5" Oct 04 04:48:28 crc kubenswrapper[4770]: I1004 04:48:28.492168 4770 scope.go:117] "RemoveContainer" containerID="cfde03b252b23508a80eee27e883f8579b863b9a2243294ca7acf7dd8f0b9717" Oct 04 04:48:28 crc kubenswrapper[4770]: I1004 04:48:28.553382 4770 scope.go:117] "RemoveContainer" containerID="1506ae294a3ccc8964a0cccf021ed83a748c273902116cc940de35ba2f635bd2" Oct 04 04:48:31 crc kubenswrapper[4770]: I1004 04:48:31.820672 4770 generic.go:334] "Generic (PLEG): container finished" podID="193a3912-2337-4217-8da9-0b67f42a6220" containerID="0b92cd5705bee2cb7dc44d7c45166bdb8831b73d134ca7f6985bb754df3f1f86" exitCode=137 Oct 04 04:48:31 crc kubenswrapper[4770]: I1004 04:48:31.820715 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69cdc756fc-p6dcj" event={"ID":"193a3912-2337-4217-8da9-0b67f42a6220","Type":"ContainerDied","Data":"0b92cd5705bee2cb7dc44d7c45166bdb8831b73d134ca7f6985bb754df3f1f86"} Oct 04 04:48:31 crc kubenswrapper[4770]: I1004 04:48:31.821082 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69cdc756fc-p6dcj" event={"ID":"193a3912-2337-4217-8da9-0b67f42a6220","Type":"ContainerDied","Data":"e870184d89ee81ae63fb1c8be41eb91f8fdd5ce7d2bf11baec03b8d48b2e80a1"} Oct 04 04:48:31 crc kubenswrapper[4770]: I1004 04:48:31.821099 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e870184d89ee81ae63fb1c8be41eb91f8fdd5ce7d2bf11baec03b8d48b2e80a1" Oct 04 04:48:31 crc kubenswrapper[4770]: I1004 04:48:31.917207 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.059676 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/193a3912-2337-4217-8da9-0b67f42a6220-logs\") pod \"193a3912-2337-4217-8da9-0b67f42a6220\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.059724 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-scripts\") pod \"193a3912-2337-4217-8da9-0b67f42a6220\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.059809 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/193a3912-2337-4217-8da9-0b67f42a6220-horizon-secret-key\") pod \"193a3912-2337-4217-8da9-0b67f42a6220\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.059857 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-config-data\") pod \"193a3912-2337-4217-8da9-0b67f42a6220\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.060126 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24pwr\" (UniqueName: \"kubernetes.io/projected/193a3912-2337-4217-8da9-0b67f42a6220-kube-api-access-24pwr\") pod \"193a3912-2337-4217-8da9-0b67f42a6220\" (UID: \"193a3912-2337-4217-8da9-0b67f42a6220\") " Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.060163 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/193a3912-2337-4217-8da9-0b67f42a6220-logs" (OuterVolumeSpecName: "logs") pod "193a3912-2337-4217-8da9-0b67f42a6220" (UID: "193a3912-2337-4217-8da9-0b67f42a6220"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.060521 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/193a3912-2337-4217-8da9-0b67f42a6220-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.067945 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/193a3912-2337-4217-8da9-0b67f42a6220-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "193a3912-2337-4217-8da9-0b67f42a6220" (UID: "193a3912-2337-4217-8da9-0b67f42a6220"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.067991 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/193a3912-2337-4217-8da9-0b67f42a6220-kube-api-access-24pwr" (OuterVolumeSpecName: "kube-api-access-24pwr") pod "193a3912-2337-4217-8da9-0b67f42a6220" (UID: "193a3912-2337-4217-8da9-0b67f42a6220"). InnerVolumeSpecName "kube-api-access-24pwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.100994 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-config-data" (OuterVolumeSpecName: "config-data") pod "193a3912-2337-4217-8da9-0b67f42a6220" (UID: "193a3912-2337-4217-8da9-0b67f42a6220"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.101216 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-scripts" (OuterVolumeSpecName: "scripts") pod "193a3912-2337-4217-8da9-0b67f42a6220" (UID: "193a3912-2337-4217-8da9-0b67f42a6220"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.161814 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.161856 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24pwr\" (UniqueName: \"kubernetes.io/projected/193a3912-2337-4217-8da9-0b67f42a6220-kube-api-access-24pwr\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.161868 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/193a3912-2337-4217-8da9-0b67f42a6220-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.161877 4770 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/193a3912-2337-4217-8da9-0b67f42a6220-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.830047 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69cdc756fc-p6dcj" Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.874663 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69cdc756fc-p6dcj"] Oct 04 04:48:32 crc kubenswrapper[4770]: I1004 04:48:32.882129 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-69cdc756fc-p6dcj"] Oct 04 04:48:33 crc kubenswrapper[4770]: I1004 04:48:33.674780 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:48:33 crc kubenswrapper[4770]: E1004 04:48:33.675214 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:48:33 crc kubenswrapper[4770]: I1004 04:48:33.693485 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="193a3912-2337-4217-8da9-0b67f42a6220" path="/var/lib/kubelet/pods/193a3912-2337-4217-8da9-0b67f42a6220/volumes" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.404071 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64f6748dff-9x2w7"] Oct 04 04:48:35 crc kubenswrapper[4770]: E1004 04:48:35.405164 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerName="extract-content" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405183 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerName="extract-content" Oct 04 04:48:35 crc kubenswrapper[4770]: E1004 04:48:35.405218 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405228 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon" Oct 04 04:48:35 crc kubenswrapper[4770]: E1004 04:48:35.405243 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerName="registry-server" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405253 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerName="registry-server" Oct 04 04:48:35 crc kubenswrapper[4770]: E1004 04:48:35.405265 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon-log" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405271 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon-log" Oct 04 04:48:35 crc kubenswrapper[4770]: E1004 04:48:35.405287 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73a9a959-f316-4699-a1b6-9ecc07204706" containerName="horizon-log" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405295 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="73a9a959-f316-4699-a1b6-9ecc07204706" containerName="horizon-log" Oct 04 04:48:35 crc kubenswrapper[4770]: E1004 04:48:35.405316 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73a9a959-f316-4699-a1b6-9ecc07204706" containerName="horizon" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405324 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="73a9a959-f316-4699-a1b6-9ecc07204706" containerName="horizon" Oct 04 04:48:35 crc kubenswrapper[4770]: E1004 04:48:35.405339 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerName="extract-utilities" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405346 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerName="extract-utilities" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405566 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="73a9a959-f316-4699-a1b6-9ecc07204706" containerName="horizon-log" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405580 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b3f1d6-060c-4d9d-931f-f08f273d556e" containerName="registry-server" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405595 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405611 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="193a3912-2337-4217-8da9-0b67f42a6220" containerName="horizon-log" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.405624 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="73a9a959-f316-4699-a1b6-9ecc07204706" containerName="horizon" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.406901 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.446835 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64f6748dff-9x2w7"] Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.544183 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/025aaf1a-ee7f-4715-8f03-dadd939d8bda-horizon-secret-key\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.544289 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025aaf1a-ee7f-4715-8f03-dadd939d8bda-logs\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.544353 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/025aaf1a-ee7f-4715-8f03-dadd939d8bda-scripts\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.544407 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/025aaf1a-ee7f-4715-8f03-dadd939d8bda-config-data\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.544568 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb7zh\" (UniqueName: \"kubernetes.io/projected/025aaf1a-ee7f-4715-8f03-dadd939d8bda-kube-api-access-mb7zh\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.646704 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025aaf1a-ee7f-4715-8f03-dadd939d8bda-logs\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.646801 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/025aaf1a-ee7f-4715-8f03-dadd939d8bda-scripts\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.646850 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/025aaf1a-ee7f-4715-8f03-dadd939d8bda-config-data\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.646954 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb7zh\" (UniqueName: \"kubernetes.io/projected/025aaf1a-ee7f-4715-8f03-dadd939d8bda-kube-api-access-mb7zh\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.647041 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/025aaf1a-ee7f-4715-8f03-dadd939d8bda-horizon-secret-key\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.647506 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025aaf1a-ee7f-4715-8f03-dadd939d8bda-logs\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.648373 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/025aaf1a-ee7f-4715-8f03-dadd939d8bda-scripts\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.648878 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/025aaf1a-ee7f-4715-8f03-dadd939d8bda-config-data\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.653332 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/025aaf1a-ee7f-4715-8f03-dadd939d8bda-horizon-secret-key\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.663948 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb7zh\" (UniqueName: \"kubernetes.io/projected/025aaf1a-ee7f-4715-8f03-dadd939d8bda-kube-api-access-mb7zh\") pod \"horizon-64f6748dff-9x2w7\" (UID: \"025aaf1a-ee7f-4715-8f03-dadd939d8bda\") " pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:35 crc kubenswrapper[4770]: I1004 04:48:35.750886 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:36 crc kubenswrapper[4770]: I1004 04:48:36.268579 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64f6748dff-9x2w7"] Oct 04 04:48:36 crc kubenswrapper[4770]: I1004 04:48:36.831217 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-c9z68"] Oct 04 04:48:36 crc kubenswrapper[4770]: I1004 04:48:36.832654 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-c9z68" Oct 04 04:48:36 crc kubenswrapper[4770]: I1004 04:48:36.840265 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-c9z68"] Oct 04 04:48:36 crc kubenswrapper[4770]: I1004 04:48:36.882963 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f6748dff-9x2w7" event={"ID":"025aaf1a-ee7f-4715-8f03-dadd939d8bda","Type":"ContainerStarted","Data":"5848c06b59b066bfc3da81e9044d20d962d49d431f71945533966f212c7f54ca"} Oct 04 04:48:36 crc kubenswrapper[4770]: I1004 04:48:36.883093 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f6748dff-9x2w7" event={"ID":"025aaf1a-ee7f-4715-8f03-dadd939d8bda","Type":"ContainerStarted","Data":"69d773271498f149ba4ef046be669d9b226e310e034a8166d881929055cac2e8"} Oct 04 04:48:36 crc kubenswrapper[4770]: I1004 04:48:36.883104 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f6748dff-9x2w7" event={"ID":"025aaf1a-ee7f-4715-8f03-dadd939d8bda","Type":"ContainerStarted","Data":"2dd542b817891032e4e11bf714e66bcfaecd853981944dcdff739139a2d73ae9"} Oct 04 04:48:36 crc kubenswrapper[4770]: I1004 04:48:36.906519 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64f6748dff-9x2w7" podStartSLOduration=1.906499335 podStartE2EDuration="1.906499335s" podCreationTimestamp="2025-10-04 04:48:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:48:36.904335909 +0000 UTC m=+6328.196345621" watchObservedRunningTime="2025-10-04 04:48:36.906499335 +0000 UTC m=+6328.198509047" Oct 04 04:48:36 crc kubenswrapper[4770]: I1004 04:48:36.978925 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq2ch\" (UniqueName: \"kubernetes.io/projected/bb09598e-d547-4375-a495-0f14fe0d0480-kube-api-access-wq2ch\") pod \"heat-db-create-c9z68\" (UID: \"bb09598e-d547-4375-a495-0f14fe0d0480\") " pod="openstack/heat-db-create-c9z68" Oct 04 04:48:37 crc kubenswrapper[4770]: I1004 04:48:37.080747 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq2ch\" (UniqueName: \"kubernetes.io/projected/bb09598e-d547-4375-a495-0f14fe0d0480-kube-api-access-wq2ch\") pod \"heat-db-create-c9z68\" (UID: \"bb09598e-d547-4375-a495-0f14fe0d0480\") " pod="openstack/heat-db-create-c9z68" Oct 04 04:48:37 crc kubenswrapper[4770]: I1004 04:48:37.105240 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq2ch\" (UniqueName: \"kubernetes.io/projected/bb09598e-d547-4375-a495-0f14fe0d0480-kube-api-access-wq2ch\") pod \"heat-db-create-c9z68\" (UID: \"bb09598e-d547-4375-a495-0f14fe0d0480\") " pod="openstack/heat-db-create-c9z68" Oct 04 04:48:37 crc kubenswrapper[4770]: I1004 04:48:37.185875 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-c9z68" Oct 04 04:48:37 crc kubenswrapper[4770]: I1004 04:48:37.685210 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-c9z68"] Oct 04 04:48:37 crc kubenswrapper[4770]: I1004 04:48:37.892748 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-c9z68" event={"ID":"bb09598e-d547-4375-a495-0f14fe0d0480","Type":"ContainerStarted","Data":"5d31d53b5f296ead598acc1a4ba3d6043db3b1063f2a424af8e2f2aead07df9f"} Oct 04 04:48:38 crc kubenswrapper[4770]: I1004 04:48:38.902864 4770 generic.go:334] "Generic (PLEG): container finished" podID="bb09598e-d547-4375-a495-0f14fe0d0480" containerID="bf97667caafea1cfca3bd6c0facac44e41462e92649bdee7d60bd2ba83014adb" exitCode=0 Oct 04 04:48:38 crc kubenswrapper[4770]: I1004 04:48:38.902979 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-c9z68" event={"ID":"bb09598e-d547-4375-a495-0f14fe0d0480","Type":"ContainerDied","Data":"bf97667caafea1cfca3bd6c0facac44e41462e92649bdee7d60bd2ba83014adb"} Oct 04 04:48:40 crc kubenswrapper[4770]: I1004 04:48:40.241234 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-c9z68" Oct 04 04:48:40 crc kubenswrapper[4770]: I1004 04:48:40.359589 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq2ch\" (UniqueName: \"kubernetes.io/projected/bb09598e-d547-4375-a495-0f14fe0d0480-kube-api-access-wq2ch\") pod \"bb09598e-d547-4375-a495-0f14fe0d0480\" (UID: \"bb09598e-d547-4375-a495-0f14fe0d0480\") " Oct 04 04:48:40 crc kubenswrapper[4770]: I1004 04:48:40.365322 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb09598e-d547-4375-a495-0f14fe0d0480-kube-api-access-wq2ch" (OuterVolumeSpecName: "kube-api-access-wq2ch") pod "bb09598e-d547-4375-a495-0f14fe0d0480" (UID: "bb09598e-d547-4375-a495-0f14fe0d0480"). InnerVolumeSpecName "kube-api-access-wq2ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:40 crc kubenswrapper[4770]: I1004 04:48:40.462910 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq2ch\" (UniqueName: \"kubernetes.io/projected/bb09598e-d547-4375-a495-0f14fe0d0480-kube-api-access-wq2ch\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:40 crc kubenswrapper[4770]: I1004 04:48:40.931073 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-c9z68" event={"ID":"bb09598e-d547-4375-a495-0f14fe0d0480","Type":"ContainerDied","Data":"5d31d53b5f296ead598acc1a4ba3d6043db3b1063f2a424af8e2f2aead07df9f"} Oct 04 04:48:40 crc kubenswrapper[4770]: I1004 04:48:40.931244 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d31d53b5f296ead598acc1a4ba3d6043db3b1063f2a424af8e2f2aead07df9f" Oct 04 04:48:40 crc kubenswrapper[4770]: I1004 04:48:40.931374 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-c9z68" Oct 04 04:48:45 crc kubenswrapper[4770]: I1004 04:48:45.751547 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:45 crc kubenswrapper[4770]: I1004 04:48:45.752148 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:48:46 crc kubenswrapper[4770]: I1004 04:48:46.677320 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:48:46 crc kubenswrapper[4770]: E1004 04:48:46.678974 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:48:46 crc kubenswrapper[4770]: I1004 04:48:46.903941 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-574f-account-create-2vw9g"] Oct 04 04:48:46 crc kubenswrapper[4770]: E1004 04:48:46.905436 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb09598e-d547-4375-a495-0f14fe0d0480" containerName="mariadb-database-create" Oct 04 04:48:46 crc kubenswrapper[4770]: I1004 04:48:46.905516 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb09598e-d547-4375-a495-0f14fe0d0480" containerName="mariadb-database-create" Oct 04 04:48:46 crc kubenswrapper[4770]: I1004 04:48:46.905799 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb09598e-d547-4375-a495-0f14fe0d0480" containerName="mariadb-database-create" Oct 04 04:48:46 crc kubenswrapper[4770]: I1004 04:48:46.906533 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-574f-account-create-2vw9g" Oct 04 04:48:46 crc kubenswrapper[4770]: I1004 04:48:46.908684 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 04 04:48:46 crc kubenswrapper[4770]: I1004 04:48:46.911268 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-574f-account-create-2vw9g"] Oct 04 04:48:47 crc kubenswrapper[4770]: I1004 04:48:47.030234 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzp5g\" (UniqueName: \"kubernetes.io/projected/ef122e98-17e7-49d3-9e07-24cf340332b5-kube-api-access-kzp5g\") pod \"heat-574f-account-create-2vw9g\" (UID: \"ef122e98-17e7-49d3-9e07-24cf340332b5\") " pod="openstack/heat-574f-account-create-2vw9g" Oct 04 04:48:47 crc kubenswrapper[4770]: I1004 04:48:47.132697 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzp5g\" (UniqueName: \"kubernetes.io/projected/ef122e98-17e7-49d3-9e07-24cf340332b5-kube-api-access-kzp5g\") pod \"heat-574f-account-create-2vw9g\" (UID: \"ef122e98-17e7-49d3-9e07-24cf340332b5\") " pod="openstack/heat-574f-account-create-2vw9g" Oct 04 04:48:47 crc kubenswrapper[4770]: I1004 04:48:47.153414 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzp5g\" (UniqueName: \"kubernetes.io/projected/ef122e98-17e7-49d3-9e07-24cf340332b5-kube-api-access-kzp5g\") pod \"heat-574f-account-create-2vw9g\" (UID: \"ef122e98-17e7-49d3-9e07-24cf340332b5\") " pod="openstack/heat-574f-account-create-2vw9g" Oct 04 04:48:47 crc kubenswrapper[4770]: I1004 04:48:47.230951 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-574f-account-create-2vw9g" Oct 04 04:48:47 crc kubenswrapper[4770]: W1004 04:48:47.688312 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef122e98_17e7_49d3_9e07_24cf340332b5.slice/crio-4f3ed16e296fbf217a6c4f058ff73e0e2b6c2b34a12530811797ee8260833963 WatchSource:0}: Error finding container 4f3ed16e296fbf217a6c4f058ff73e0e2b6c2b34a12530811797ee8260833963: Status 404 returned error can't find the container with id 4f3ed16e296fbf217a6c4f058ff73e0e2b6c2b34a12530811797ee8260833963 Oct 04 04:48:47 crc kubenswrapper[4770]: I1004 04:48:47.695404 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-574f-account-create-2vw9g"] Oct 04 04:48:48 crc kubenswrapper[4770]: I1004 04:48:48.000068 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-574f-account-create-2vw9g" event={"ID":"ef122e98-17e7-49d3-9e07-24cf340332b5","Type":"ContainerStarted","Data":"f9fe0391b3d07589167f36be94e5b4f1f4953d5b7029d383994b41bf2d87f1d7"} Oct 04 04:48:48 crc kubenswrapper[4770]: I1004 04:48:48.000361 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-574f-account-create-2vw9g" event={"ID":"ef122e98-17e7-49d3-9e07-24cf340332b5","Type":"ContainerStarted","Data":"4f3ed16e296fbf217a6c4f058ff73e0e2b6c2b34a12530811797ee8260833963"} Oct 04 04:48:49 crc kubenswrapper[4770]: I1004 04:48:49.011467 4770 generic.go:334] "Generic (PLEG): container finished" podID="ef122e98-17e7-49d3-9e07-24cf340332b5" containerID="f9fe0391b3d07589167f36be94e5b4f1f4953d5b7029d383994b41bf2d87f1d7" exitCode=0 Oct 04 04:48:49 crc kubenswrapper[4770]: I1004 04:48:49.011519 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-574f-account-create-2vw9g" event={"ID":"ef122e98-17e7-49d3-9e07-24cf340332b5","Type":"ContainerDied","Data":"f9fe0391b3d07589167f36be94e5b4f1f4953d5b7029d383994b41bf2d87f1d7"} Oct 04 04:48:49 crc kubenswrapper[4770]: I1004 04:48:49.412598 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-574f-account-create-2vw9g" Oct 04 04:48:49 crc kubenswrapper[4770]: I1004 04:48:49.593997 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzp5g\" (UniqueName: \"kubernetes.io/projected/ef122e98-17e7-49d3-9e07-24cf340332b5-kube-api-access-kzp5g\") pod \"ef122e98-17e7-49d3-9e07-24cf340332b5\" (UID: \"ef122e98-17e7-49d3-9e07-24cf340332b5\") " Oct 04 04:48:49 crc kubenswrapper[4770]: I1004 04:48:49.608898 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef122e98-17e7-49d3-9e07-24cf340332b5-kube-api-access-kzp5g" (OuterVolumeSpecName: "kube-api-access-kzp5g") pod "ef122e98-17e7-49d3-9e07-24cf340332b5" (UID: "ef122e98-17e7-49d3-9e07-24cf340332b5"). InnerVolumeSpecName "kube-api-access-kzp5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:49 crc kubenswrapper[4770]: I1004 04:48:49.698667 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzp5g\" (UniqueName: \"kubernetes.io/projected/ef122e98-17e7-49d3-9e07-24cf340332b5-kube-api-access-kzp5g\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:50 crc kubenswrapper[4770]: I1004 04:48:50.025885 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-574f-account-create-2vw9g" event={"ID":"ef122e98-17e7-49d3-9e07-24cf340332b5","Type":"ContainerDied","Data":"4f3ed16e296fbf217a6c4f058ff73e0e2b6c2b34a12530811797ee8260833963"} Oct 04 04:48:50 crc kubenswrapper[4770]: I1004 04:48:50.026267 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f3ed16e296fbf217a6c4f058ff73e0e2b6c2b34a12530811797ee8260833963" Oct 04 04:48:50 crc kubenswrapper[4770]: I1004 04:48:50.026331 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-574f-account-create-2vw9g" Oct 04 04:48:51 crc kubenswrapper[4770]: I1004 04:48:51.978681 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-fjp9q"] Oct 04 04:48:51 crc kubenswrapper[4770]: E1004 04:48:51.979520 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef122e98-17e7-49d3-9e07-24cf340332b5" containerName="mariadb-account-create" Oct 04 04:48:51 crc kubenswrapper[4770]: I1004 04:48:51.979538 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef122e98-17e7-49d3-9e07-24cf340332b5" containerName="mariadb-account-create" Oct 04 04:48:51 crc kubenswrapper[4770]: I1004 04:48:51.979798 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef122e98-17e7-49d3-9e07-24cf340332b5" containerName="mariadb-account-create" Oct 04 04:48:51 crc kubenswrapper[4770]: I1004 04:48:51.980638 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:51 crc kubenswrapper[4770]: I1004 04:48:51.983690 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-ks9bm" Oct 04 04:48:51 crc kubenswrapper[4770]: I1004 04:48:51.984044 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 04 04:48:51 crc kubenswrapper[4770]: I1004 04:48:51.989927 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-fjp9q"] Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.153714 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-combined-ca-bundle\") pod \"heat-db-sync-fjp9q\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.154104 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csklh\" (UniqueName: \"kubernetes.io/projected/514cf20d-92df-4de4-a2b7-5607902ba081-kube-api-access-csklh\") pod \"heat-db-sync-fjp9q\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.154259 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-config-data\") pod \"heat-db-sync-fjp9q\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.256209 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csklh\" (UniqueName: \"kubernetes.io/projected/514cf20d-92df-4de4-a2b7-5607902ba081-kube-api-access-csklh\") pod \"heat-db-sync-fjp9q\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.256514 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-config-data\") pod \"heat-db-sync-fjp9q\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.256680 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-combined-ca-bundle\") pod \"heat-db-sync-fjp9q\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.262069 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-combined-ca-bundle\") pod \"heat-db-sync-fjp9q\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.262249 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-config-data\") pod \"heat-db-sync-fjp9q\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.275577 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csklh\" (UniqueName: \"kubernetes.io/projected/514cf20d-92df-4de4-a2b7-5607902ba081-kube-api-access-csklh\") pod \"heat-db-sync-fjp9q\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.313481 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fjp9q" Oct 04 04:48:52 crc kubenswrapper[4770]: I1004 04:48:52.751321 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-fjp9q"] Oct 04 04:48:53 crc kubenswrapper[4770]: I1004 04:48:53.058789 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fjp9q" event={"ID":"514cf20d-92df-4de4-a2b7-5607902ba081","Type":"ContainerStarted","Data":"cb9b1c8b8b21d5518f67b7fc9989ea59422fabeeaf31169f1ab8a5ead72db63a"} Oct 04 04:48:55 crc kubenswrapper[4770]: I1004 04:48:55.755115 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f6748dff-9x2w7" podUID="025aaf1a-ee7f-4715-8f03-dadd939d8bda" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.118:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.118:8080: connect: connection refused" Oct 04 04:48:56 crc kubenswrapper[4770]: I1004 04:48:56.075475 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-g7gcl"] Oct 04 04:48:56 crc kubenswrapper[4770]: I1004 04:48:56.086708 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-g7gcl"] Oct 04 04:48:57 crc kubenswrapper[4770]: I1004 04:48:57.690305 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b33454c9-e1a4-4239-a473-5b47223d090d" path="/var/lib/kubelet/pods/b33454c9-e1a4-4239-a473-5b47223d090d/volumes" Oct 04 04:48:58 crc kubenswrapper[4770]: I1004 04:48:58.674042 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:48:58 crc kubenswrapper[4770]: E1004 04:48:58.674313 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:49:03 crc kubenswrapper[4770]: I1004 04:49:03.165338 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fjp9q" event={"ID":"514cf20d-92df-4de4-a2b7-5607902ba081","Type":"ContainerStarted","Data":"f1bcb7e4fe0bfa6f8ffc1cdc136c15c6fc3c41c7a30d8f60281db6ac74f086de"} Oct 04 04:49:03 crc kubenswrapper[4770]: I1004 04:49:03.188252 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-fjp9q" podStartSLOduration=2.495034884 podStartE2EDuration="12.188230782s" podCreationTimestamp="2025-10-04 04:48:51 +0000 UTC" firstStartedPulling="2025-10-04 04:48:52.763094739 +0000 UTC m=+6344.055104451" lastFinishedPulling="2025-10-04 04:49:02.456290637 +0000 UTC m=+6353.748300349" observedRunningTime="2025-10-04 04:49:03.178386274 +0000 UTC m=+6354.470395996" watchObservedRunningTime="2025-10-04 04:49:03.188230782 +0000 UTC m=+6354.480240504" Oct 04 04:49:05 crc kubenswrapper[4770]: I1004 04:49:05.183160 4770 generic.go:334] "Generic (PLEG): container finished" podID="514cf20d-92df-4de4-a2b7-5607902ba081" containerID="f1bcb7e4fe0bfa6f8ffc1cdc136c15c6fc3c41c7a30d8f60281db6ac74f086de" exitCode=0 Oct 04 04:49:05 crc kubenswrapper[4770]: I1004 04:49:05.183222 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fjp9q" event={"ID":"514cf20d-92df-4de4-a2b7-5607902ba081","Type":"ContainerDied","Data":"f1bcb7e4fe0bfa6f8ffc1cdc136c15c6fc3c41c7a30d8f60281db6ac74f086de"} Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.567528 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fjp9q" Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.760440 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-combined-ca-bundle\") pod \"514cf20d-92df-4de4-a2b7-5607902ba081\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.760524 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csklh\" (UniqueName: \"kubernetes.io/projected/514cf20d-92df-4de4-a2b7-5607902ba081-kube-api-access-csklh\") pod \"514cf20d-92df-4de4-a2b7-5607902ba081\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.760727 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-config-data\") pod \"514cf20d-92df-4de4-a2b7-5607902ba081\" (UID: \"514cf20d-92df-4de4-a2b7-5607902ba081\") " Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.773754 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/514cf20d-92df-4de4-a2b7-5607902ba081-kube-api-access-csklh" (OuterVolumeSpecName: "kube-api-access-csklh") pod "514cf20d-92df-4de4-a2b7-5607902ba081" (UID: "514cf20d-92df-4de4-a2b7-5607902ba081"). InnerVolumeSpecName "kube-api-access-csklh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.788162 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "514cf20d-92df-4de4-a2b7-5607902ba081" (UID: "514cf20d-92df-4de4-a2b7-5607902ba081"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.847579 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-config-data" (OuterVolumeSpecName: "config-data") pod "514cf20d-92df-4de4-a2b7-5607902ba081" (UID: "514cf20d-92df-4de4-a2b7-5607902ba081"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.862871 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csklh\" (UniqueName: \"kubernetes.io/projected/514cf20d-92df-4de4-a2b7-5607902ba081-kube-api-access-csklh\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.862906 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:06 crc kubenswrapper[4770]: I1004 04:49:06.862916 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/514cf20d-92df-4de4-a2b7-5607902ba081-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:07 crc kubenswrapper[4770]: I1004 04:49:07.042632 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-fdc9-account-create-j7fft"] Oct 04 04:49:07 crc kubenswrapper[4770]: I1004 04:49:07.051292 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-fdc9-account-create-j7fft"] Oct 04 04:49:07 crc kubenswrapper[4770]: I1004 04:49:07.202661 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fjp9q" event={"ID":"514cf20d-92df-4de4-a2b7-5607902ba081","Type":"ContainerDied","Data":"cb9b1c8b8b21d5518f67b7fc9989ea59422fabeeaf31169f1ab8a5ead72db63a"} Oct 04 04:49:07 crc kubenswrapper[4770]: I1004 04:49:07.202869 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb9b1c8b8b21d5518f67b7fc9989ea59422fabeeaf31169f1ab8a5ead72db63a" Oct 04 04:49:07 crc kubenswrapper[4770]: I1004 04:49:07.203020 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fjp9q" Oct 04 04:49:07 crc kubenswrapper[4770]: I1004 04:49:07.688619 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b250070c-a70b-472c-9800-3a3797d57101" path="/var/lib/kubelet/pods/b250070c-a70b-472c-9800-3a3797d57101/volumes" Oct 04 04:49:07 crc kubenswrapper[4770]: I1004 04:49:07.703888 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.499381 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-764bcdf645-85424"] Oct 04 04:49:08 crc kubenswrapper[4770]: E1004 04:49:08.499931 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="514cf20d-92df-4de4-a2b7-5607902ba081" containerName="heat-db-sync" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.499956 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="514cf20d-92df-4de4-a2b7-5607902ba081" containerName="heat-db-sync" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.500233 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="514cf20d-92df-4de4-a2b7-5607902ba081" containerName="heat-db-sync" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.501348 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.503156 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-ks9bm" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.503413 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.515670 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-764bcdf645-85424"] Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.521297 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.600768 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4q6r\" (UniqueName: \"kubernetes.io/projected/b5a4f316-6949-4bb8-9192-30adda4c945e-kube-api-access-f4q6r\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.600910 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5a4f316-6949-4bb8-9192-30adda4c945e-config-data-custom\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.600982 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a4f316-6949-4bb8-9192-30adda4c945e-combined-ca-bundle\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.601245 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a4f316-6949-4bb8-9192-30adda4c945e-config-data\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.639980 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-755fc856cf-4s9v6"] Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.643256 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.648325 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.658715 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-755fc856cf-4s9v6"] Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.692616 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5b95c7f4c5-559tc"] Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.693923 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.701302 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.702669 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a4f316-6949-4bb8-9192-30adda4c945e-config-data\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.702785 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4q6r\" (UniqueName: \"kubernetes.io/projected/b5a4f316-6949-4bb8-9192-30adda4c945e-kube-api-access-f4q6r\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.702881 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5a4f316-6949-4bb8-9192-30adda4c945e-config-data-custom\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.702949 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a4f316-6949-4bb8-9192-30adda4c945e-combined-ca-bundle\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.711673 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5a4f316-6949-4bb8-9192-30adda4c945e-config-data-custom\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.713693 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a4f316-6949-4bb8-9192-30adda4c945e-config-data\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.729958 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a4f316-6949-4bb8-9192-30adda4c945e-combined-ca-bundle\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.733122 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b95c7f4c5-559tc"] Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.746633 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4q6r\" (UniqueName: \"kubernetes.io/projected/b5a4f316-6949-4bb8-9192-30adda4c945e-kube-api-access-f4q6r\") pod \"heat-engine-764bcdf645-85424\" (UID: \"b5a4f316-6949-4bb8-9192-30adda4c945e\") " pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.804413 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7f9w\" (UniqueName: \"kubernetes.io/projected/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-kube-api-access-b7f9w\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.804492 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-config-data-custom\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.804626 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-combined-ca-bundle\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.804668 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-config-data-custom\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.804768 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-config-data\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.804792 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-config-data\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.804850 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scwqk\" (UniqueName: \"kubernetes.io/projected/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-kube-api-access-scwqk\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.805718 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-combined-ca-bundle\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.822368 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.907183 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-combined-ca-bundle\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.907258 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-config-data-custom\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.907349 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-config-data\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.907382 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-config-data\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.907419 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scwqk\" (UniqueName: \"kubernetes.io/projected/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-kube-api-access-scwqk\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.907461 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-combined-ca-bundle\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.907516 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7f9w\" (UniqueName: \"kubernetes.io/projected/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-kube-api-access-b7f9w\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.907569 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-config-data-custom\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.919398 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-combined-ca-bundle\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.923782 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-config-data-custom\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.924465 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-config-data-custom\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.924575 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-combined-ca-bundle\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.925327 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-config-data\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.932319 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-config-data\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.943032 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scwqk\" (UniqueName: \"kubernetes.io/projected/dd09edfd-2fa6-4516-b3d6-8937cab9ff76-kube-api-access-scwqk\") pod \"heat-cfnapi-5b95c7f4c5-559tc\" (UID: \"dd09edfd-2fa6-4516-b3d6-8937cab9ff76\") " pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.943135 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7f9w\" (UniqueName: \"kubernetes.io/projected/58ed5893-3dd0-42bf-b6ea-f2ff90808c98-kube-api-access-b7f9w\") pod \"heat-api-755fc856cf-4s9v6\" (UID: \"58ed5893-3dd0-42bf-b6ea-f2ff90808c98\") " pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:08 crc kubenswrapper[4770]: I1004 04:49:08.975606 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:09 crc kubenswrapper[4770]: I1004 04:49:09.125990 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:09 crc kubenswrapper[4770]: I1004 04:49:09.429865 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-764bcdf645-85424"] Oct 04 04:49:09 crc kubenswrapper[4770]: I1004 04:49:09.658592 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-755fc856cf-4s9v6"] Oct 04 04:49:09 crc kubenswrapper[4770]: I1004 04:49:09.856638 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b95c7f4c5-559tc"] Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.126074 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-64f6748dff-9x2w7" Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.188832 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58946648d7-rhdw9"] Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.189113 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-58946648d7-rhdw9" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon-log" containerID="cri-o://497f70a99994b7812505d59ac409e22473748a32eb26c4d48c5fc7a115fb75e1" gracePeriod=30 Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.189259 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-58946648d7-rhdw9" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon" containerID="cri-o://77b68824627eefb541868b796c23f6f96c21155c153751c88765a42c0fb63581" gracePeriod=30 Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.235404 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-764bcdf645-85424" event={"ID":"b5a4f316-6949-4bb8-9192-30adda4c945e","Type":"ContainerStarted","Data":"d4bad05de637f6421a0d7d1d565846685906f190b7c9506ba5bed69956ce64cb"} Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.235457 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-764bcdf645-85424" event={"ID":"b5a4f316-6949-4bb8-9192-30adda4c945e","Type":"ContainerStarted","Data":"46905d08da2925790ff98b44dd243805f196a5730e2730f2f6a2763c91577943"} Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.235773 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.236972 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-755fc856cf-4s9v6" event={"ID":"58ed5893-3dd0-42bf-b6ea-f2ff90808c98","Type":"ContainerStarted","Data":"70f232b98b2b3a974f0dc5a007df85a3c8071378136b017736435cf1f2ab9956"} Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.242140 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" event={"ID":"dd09edfd-2fa6-4516-b3d6-8937cab9ff76","Type":"ContainerStarted","Data":"b4a5ecb77e8282d793717963e1506e44654a85e08b0b0c89273c38afe3c21cd8"} Oct 04 04:49:10 crc kubenswrapper[4770]: I1004 04:49:10.267606 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-764bcdf645-85424" podStartSLOduration=2.267583834 podStartE2EDuration="2.267583834s" podCreationTimestamp="2025-10-04 04:49:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:49:10.260018717 +0000 UTC m=+6361.552028429" watchObservedRunningTime="2025-10-04 04:49:10.267583834 +0000 UTC m=+6361.559593556" Oct 04 04:49:12 crc kubenswrapper[4770]: I1004 04:49:12.269045 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-755fc856cf-4s9v6" event={"ID":"58ed5893-3dd0-42bf-b6ea-f2ff90808c98","Type":"ContainerStarted","Data":"a6461d2297ae69ec44b4a7377f0b804ca921a0a0f26e3620f4b5c7c6e72b9dd6"} Oct 04 04:49:12 crc kubenswrapper[4770]: I1004 04:49:12.269738 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:12 crc kubenswrapper[4770]: I1004 04:49:12.298186 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-755fc856cf-4s9v6" podStartSLOduration=2.818254366 podStartE2EDuration="4.298163864s" podCreationTimestamp="2025-10-04 04:49:08 +0000 UTC" firstStartedPulling="2025-10-04 04:49:09.681751692 +0000 UTC m=+6360.973761404" lastFinishedPulling="2025-10-04 04:49:11.16166119 +0000 UTC m=+6362.453670902" observedRunningTime="2025-10-04 04:49:12.292325562 +0000 UTC m=+6363.584335274" watchObservedRunningTime="2025-10-04 04:49:12.298163864 +0000 UTC m=+6363.590173576" Oct 04 04:49:13 crc kubenswrapper[4770]: I1004 04:49:13.673536 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:49:13 crc kubenswrapper[4770]: E1004 04:49:13.674318 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:49:14 crc kubenswrapper[4770]: I1004 04:49:14.307315 4770 generic.go:334] "Generic (PLEG): container finished" podID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerID="77b68824627eefb541868b796c23f6f96c21155c153751c88765a42c0fb63581" exitCode=0 Oct 04 04:49:14 crc kubenswrapper[4770]: I1004 04:49:14.307449 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58946648d7-rhdw9" event={"ID":"e8d763f0-83cf-4dfd-975e-186f6149274e","Type":"ContainerDied","Data":"77b68824627eefb541868b796c23f6f96c21155c153751c88765a42c0fb63581"} Oct 04 04:49:16 crc kubenswrapper[4770]: I1004 04:49:16.326593 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" event={"ID":"dd09edfd-2fa6-4516-b3d6-8937cab9ff76","Type":"ContainerStarted","Data":"a4647360cfbe0157f621cb116ca5004bcee5aa52eca2a3919db1ea7eebcb6fb7"} Oct 04 04:49:16 crc kubenswrapper[4770]: I1004 04:49:16.327054 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:16 crc kubenswrapper[4770]: I1004 04:49:16.345843 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" podStartSLOduration=2.331172688 podStartE2EDuration="8.345824033s" podCreationTimestamp="2025-10-04 04:49:08 +0000 UTC" firstStartedPulling="2025-10-04 04:49:09.866164775 +0000 UTC m=+6361.158174487" lastFinishedPulling="2025-10-04 04:49:15.88081612 +0000 UTC m=+6367.172825832" observedRunningTime="2025-10-04 04:49:16.339939799 +0000 UTC m=+6367.631949511" watchObservedRunningTime="2025-10-04 04:49:16.345824033 +0000 UTC m=+6367.637833745" Oct 04 04:49:17 crc kubenswrapper[4770]: I1004 04:49:17.455862 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-58946648d7-rhdw9" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 04 04:49:20 crc kubenswrapper[4770]: I1004 04:49:20.321272 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-755fc856cf-4s9v6" Oct 04 04:49:25 crc kubenswrapper[4770]: I1004 04:49:25.498656 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5b95c7f4c5-559tc" Oct 04 04:49:25 crc kubenswrapper[4770]: I1004 04:49:25.674322 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:49:25 crc kubenswrapper[4770]: E1004 04:49:25.674800 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:49:27 crc kubenswrapper[4770]: I1004 04:49:27.456095 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-58946648d7-rhdw9" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 04 04:49:28 crc kubenswrapper[4770]: I1004 04:49:28.807790 4770 scope.go:117] "RemoveContainer" containerID="29ffdc2a9210f2df9a291930a656a35612c13f93ab4ced3a39575d1362c0cece" Oct 04 04:49:28 crc kubenswrapper[4770]: I1004 04:49:28.828147 4770 scope.go:117] "RemoveContainer" containerID="a97e14ae27fb18684c6f130fd5e5c7568cf392c4f90c5a078ffdad08e10fe1c1" Oct 04 04:49:28 crc kubenswrapper[4770]: I1004 04:49:28.857417 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-764bcdf645-85424" Oct 04 04:49:37 crc kubenswrapper[4770]: I1004 04:49:37.456829 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-58946648d7-rhdw9" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 04 04:49:37 crc kubenswrapper[4770]: I1004 04:49:37.457786 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.341847 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr"] Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.345786 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.347461 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.350757 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr"] Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.459396 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.459511 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb8mz\" (UniqueName: \"kubernetes.io/projected/4129621b-ab45-4ed3-b43f-590caee83445-kube-api-access-sb8mz\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.459876 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.562021 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb8mz\" (UniqueName: \"kubernetes.io/projected/4129621b-ab45-4ed3-b43f-590caee83445-kube-api-access-sb8mz\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.562151 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.562294 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.562785 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.562827 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.583414 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb8mz\" (UniqueName: \"kubernetes.io/projected/4129621b-ab45-4ed3-b43f-590caee83445-kube-api-access-sb8mz\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:38 crc kubenswrapper[4770]: I1004 04:49:38.675582 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:39 crc kubenswrapper[4770]: I1004 04:49:39.253362 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr"] Oct 04 04:49:39 crc kubenswrapper[4770]: W1004 04:49:39.413830 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4129621b_ab45_4ed3_b43f_590caee83445.slice/crio-1fb6a742a6643adb54c14d7f17d165dfd1c77e5621913ce2fcc1f94b1b21fe6f WatchSource:0}: Error finding container 1fb6a742a6643adb54c14d7f17d165dfd1c77e5621913ce2fcc1f94b1b21fe6f: Status 404 returned error can't find the container with id 1fb6a742a6643adb54c14d7f17d165dfd1c77e5621913ce2fcc1f94b1b21fe6f Oct 04 04:49:39 crc kubenswrapper[4770]: I1004 04:49:39.603162 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" event={"ID":"4129621b-ab45-4ed3-b43f-590caee83445","Type":"ContainerStarted","Data":"1fb6a742a6643adb54c14d7f17d165dfd1c77e5621913ce2fcc1f94b1b21fe6f"} Oct 04 04:49:40 crc kubenswrapper[4770]: I1004 04:49:40.632081 4770 generic.go:334] "Generic (PLEG): container finished" podID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerID="497f70a99994b7812505d59ac409e22473748a32eb26c4d48c5fc7a115fb75e1" exitCode=137 Oct 04 04:49:40 crc kubenswrapper[4770]: I1004 04:49:40.632292 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58946648d7-rhdw9" event={"ID":"e8d763f0-83cf-4dfd-975e-186f6149274e","Type":"ContainerDied","Data":"497f70a99994b7812505d59ac409e22473748a32eb26c4d48c5fc7a115fb75e1"} Oct 04 04:49:40 crc kubenswrapper[4770]: I1004 04:49:40.635403 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" event={"ID":"4129621b-ab45-4ed3-b43f-590caee83445","Type":"ContainerStarted","Data":"31effaf3a471408083a6f1f8721abb4d36d9df93b552589694ad5f3d8c8e3507"} Oct 04 04:49:40 crc kubenswrapper[4770]: I1004 04:49:40.673938 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.346630 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.422558 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8d763f0-83cf-4dfd-975e-186f6149274e-logs\") pod \"e8d763f0-83cf-4dfd-975e-186f6149274e\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.422752 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-config-data\") pod \"e8d763f0-83cf-4dfd-975e-186f6149274e\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.422845 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8d763f0-83cf-4dfd-975e-186f6149274e-horizon-secret-key\") pod \"e8d763f0-83cf-4dfd-975e-186f6149274e\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.422902 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbtzv\" (UniqueName: \"kubernetes.io/projected/e8d763f0-83cf-4dfd-975e-186f6149274e-kube-api-access-nbtzv\") pod \"e8d763f0-83cf-4dfd-975e-186f6149274e\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.422946 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-scripts\") pod \"e8d763f0-83cf-4dfd-975e-186f6149274e\" (UID: \"e8d763f0-83cf-4dfd-975e-186f6149274e\") " Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.423891 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8d763f0-83cf-4dfd-975e-186f6149274e-logs" (OuterVolumeSpecName: "logs") pod "e8d763f0-83cf-4dfd-975e-186f6149274e" (UID: "e8d763f0-83cf-4dfd-975e-186f6149274e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.428573 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8d763f0-83cf-4dfd-975e-186f6149274e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e8d763f0-83cf-4dfd-975e-186f6149274e" (UID: "e8d763f0-83cf-4dfd-975e-186f6149274e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.428617 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8d763f0-83cf-4dfd-975e-186f6149274e-kube-api-access-nbtzv" (OuterVolumeSpecName: "kube-api-access-nbtzv") pod "e8d763f0-83cf-4dfd-975e-186f6149274e" (UID: "e8d763f0-83cf-4dfd-975e-186f6149274e"). InnerVolumeSpecName "kube-api-access-nbtzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.449570 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-config-data" (OuterVolumeSpecName: "config-data") pod "e8d763f0-83cf-4dfd-975e-186f6149274e" (UID: "e8d763f0-83cf-4dfd-975e-186f6149274e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.451399 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-scripts" (OuterVolumeSpecName: "scripts") pod "e8d763f0-83cf-4dfd-975e-186f6149274e" (UID: "e8d763f0-83cf-4dfd-975e-186f6149274e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.525834 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8d763f0-83cf-4dfd-975e-186f6149274e-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.525874 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.525886 4770 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8d763f0-83cf-4dfd-975e-186f6149274e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.525897 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbtzv\" (UniqueName: \"kubernetes.io/projected/e8d763f0-83cf-4dfd-975e-186f6149274e-kube-api-access-nbtzv\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.525908 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8d763f0-83cf-4dfd-975e-186f6149274e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.645179 4770 generic.go:334] "Generic (PLEG): container finished" podID="4129621b-ab45-4ed3-b43f-590caee83445" containerID="31effaf3a471408083a6f1f8721abb4d36d9df93b552589694ad5f3d8c8e3507" exitCode=0 Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.645230 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" event={"ID":"4129621b-ab45-4ed3-b43f-590caee83445","Type":"ContainerDied","Data":"31effaf3a471408083a6f1f8721abb4d36d9df93b552589694ad5f3d8c8e3507"} Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.649668 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"284e8fec1a55dcb36285518859a2a4bc58d3161ca3219d7437f89da44e6444b6"} Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.651965 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58946648d7-rhdw9" event={"ID":"e8d763f0-83cf-4dfd-975e-186f6149274e","Type":"ContainerDied","Data":"750ad3db6dbd6a2f90bfb7076f505d94aef03b74ab2f5e8050cd710ca31832a6"} Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.652000 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58946648d7-rhdw9" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.652022 4770 scope.go:117] "RemoveContainer" containerID="77b68824627eefb541868b796c23f6f96c21155c153751c88765a42c0fb63581" Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.716885 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58946648d7-rhdw9"] Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.725936 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-58946648d7-rhdw9"] Oct 04 04:49:41 crc kubenswrapper[4770]: I1004 04:49:41.847738 4770 scope.go:117] "RemoveContainer" containerID="497f70a99994b7812505d59ac409e22473748a32eb26c4d48c5fc7a115fb75e1" Oct 04 04:49:43 crc kubenswrapper[4770]: I1004 04:49:43.705539 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" path="/var/lib/kubelet/pods/e8d763f0-83cf-4dfd-975e-186f6149274e/volumes" Oct 04 04:49:45 crc kubenswrapper[4770]: I1004 04:49:45.044335 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-pgbb9"] Oct 04 04:49:45 crc kubenswrapper[4770]: I1004 04:49:45.053178 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-pgbb9"] Oct 04 04:49:45 crc kubenswrapper[4770]: I1004 04:49:45.688276 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5f22307-4c2c-4c2b-82d6-6e323d4995e4" path="/var/lib/kubelet/pods/e5f22307-4c2c-4c2b-82d6-6e323d4995e4/volumes" Oct 04 04:49:46 crc kubenswrapper[4770]: I1004 04:49:46.707912 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" event={"ID":"4129621b-ab45-4ed3-b43f-590caee83445","Type":"ContainerStarted","Data":"6c0f6e49bf779ff3840f45971c46aad3e6fa1f9c134e564b009ec5a4cc651fd7"} Oct 04 04:49:48 crc kubenswrapper[4770]: I1004 04:49:48.734553 4770 generic.go:334] "Generic (PLEG): container finished" podID="4129621b-ab45-4ed3-b43f-590caee83445" containerID="6c0f6e49bf779ff3840f45971c46aad3e6fa1f9c134e564b009ec5a4cc651fd7" exitCode=0 Oct 04 04:49:48 crc kubenswrapper[4770]: I1004 04:49:48.734650 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" event={"ID":"4129621b-ab45-4ed3-b43f-590caee83445","Type":"ContainerDied","Data":"6c0f6e49bf779ff3840f45971c46aad3e6fa1f9c134e564b009ec5a4cc651fd7"} Oct 04 04:49:49 crc kubenswrapper[4770]: I1004 04:49:49.747668 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" event={"ID":"4129621b-ab45-4ed3-b43f-590caee83445","Type":"ContainerStarted","Data":"fb4a1e3f13ae3a4598fe8dd99c9bca0b225e74a9aa0d892840a8edff2e9873dd"} Oct 04 04:49:51 crc kubenswrapper[4770]: I1004 04:49:51.772779 4770 generic.go:334] "Generic (PLEG): container finished" podID="4129621b-ab45-4ed3-b43f-590caee83445" containerID="fb4a1e3f13ae3a4598fe8dd99c9bca0b225e74a9aa0d892840a8edff2e9873dd" exitCode=0 Oct 04 04:49:51 crc kubenswrapper[4770]: I1004 04:49:51.772905 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" event={"ID":"4129621b-ab45-4ed3-b43f-590caee83445","Type":"ContainerDied","Data":"fb4a1e3f13ae3a4598fe8dd99c9bca0b225e74a9aa0d892840a8edff2e9873dd"} Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.416757 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.497348 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-bundle\") pod \"4129621b-ab45-4ed3-b43f-590caee83445\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.497502 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb8mz\" (UniqueName: \"kubernetes.io/projected/4129621b-ab45-4ed3-b43f-590caee83445-kube-api-access-sb8mz\") pod \"4129621b-ab45-4ed3-b43f-590caee83445\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.497555 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-util\") pod \"4129621b-ab45-4ed3-b43f-590caee83445\" (UID: \"4129621b-ab45-4ed3-b43f-590caee83445\") " Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.500519 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-bundle" (OuterVolumeSpecName: "bundle") pod "4129621b-ab45-4ed3-b43f-590caee83445" (UID: "4129621b-ab45-4ed3-b43f-590caee83445"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.505343 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4129621b-ab45-4ed3-b43f-590caee83445-kube-api-access-sb8mz" (OuterVolumeSpecName: "kube-api-access-sb8mz") pod "4129621b-ab45-4ed3-b43f-590caee83445" (UID: "4129621b-ab45-4ed3-b43f-590caee83445"). InnerVolumeSpecName "kube-api-access-sb8mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.517831 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-util" (OuterVolumeSpecName: "util") pod "4129621b-ab45-4ed3-b43f-590caee83445" (UID: "4129621b-ab45-4ed3-b43f-590caee83445"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.601575 4770 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.601631 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb8mz\" (UniqueName: \"kubernetes.io/projected/4129621b-ab45-4ed3-b43f-590caee83445-kube-api-access-sb8mz\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.601643 4770 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4129621b-ab45-4ed3-b43f-590caee83445-util\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.791819 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" event={"ID":"4129621b-ab45-4ed3-b43f-590caee83445","Type":"ContainerDied","Data":"1fb6a742a6643adb54c14d7f17d165dfd1c77e5621913ce2fcc1f94b1b21fe6f"} Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.792226 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fb6a742a6643adb54c14d7f17d165dfd1c77e5621913ce2fcc1f94b1b21fe6f" Oct 04 04:49:53 crc kubenswrapper[4770]: I1004 04:49:53.791894 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.465041 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv"] Oct 04 04:50:06 crc kubenswrapper[4770]: E1004 04:50:06.466036 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4129621b-ab45-4ed3-b43f-590caee83445" containerName="extract" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.466049 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4129621b-ab45-4ed3-b43f-590caee83445" containerName="extract" Oct 04 04:50:06 crc kubenswrapper[4770]: E1004 04:50:06.466088 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4129621b-ab45-4ed3-b43f-590caee83445" containerName="util" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.466094 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4129621b-ab45-4ed3-b43f-590caee83445" containerName="util" Oct 04 04:50:06 crc kubenswrapper[4770]: E1004 04:50:06.466100 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4129621b-ab45-4ed3-b43f-590caee83445" containerName="pull" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.466107 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="4129621b-ab45-4ed3-b43f-590caee83445" containerName="pull" Oct 04 04:50:06 crc kubenswrapper[4770]: E1004 04:50:06.466122 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.466130 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon" Oct 04 04:50:06 crc kubenswrapper[4770]: E1004 04:50:06.466146 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon-log" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.466153 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon-log" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.466549 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon-log" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.466567 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="4129621b-ab45-4ed3-b43f-590caee83445" containerName="extract" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.466586 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8d763f0-83cf-4dfd-975e-186f6149274e" containerName="horizon" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.467636 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.470891 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.471167 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-d59tg" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.474105 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.528637 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv"] Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.609603 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpk5f\" (UniqueName: \"kubernetes.io/projected/b55bd1e1-8afd-42a0-b180-c91c3f66fb4d-kube-api-access-dpk5f\") pod \"obo-prometheus-operator-7c8cf85677-d5cfv\" (UID: \"b55bd1e1-8afd-42a0-b180-c91c3f66fb4d\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.648315 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb"] Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.649538 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.658675 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-g575t" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.664624 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv"] Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.683345 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.712266 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb"] Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.712722 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.712808 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b3ecc0f-c11f-4572-8303-65ba8f955b03-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb\" (UID: \"6b3ecc0f-c11f-4572-8303-65ba8f955b03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.712911 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpk5f\" (UniqueName: \"kubernetes.io/projected/b55bd1e1-8afd-42a0-b180-c91c3f66fb4d-kube-api-access-dpk5f\") pod \"obo-prometheus-operator-7c8cf85677-d5cfv\" (UID: \"b55bd1e1-8afd-42a0-b180-c91c3f66fb4d\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.712936 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b3ecc0f-c11f-4572-8303-65ba8f955b03-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb\" (UID: \"6b3ecc0f-c11f-4572-8303-65ba8f955b03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.725704 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv"] Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.767891 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpk5f\" (UniqueName: \"kubernetes.io/projected/b55bd1e1-8afd-42a0-b180-c91c3f66fb4d-kube-api-access-dpk5f\") pod \"obo-prometheus-operator-7c8cf85677-d5cfv\" (UID: \"b55bd1e1-8afd-42a0-b180-c91c3f66fb4d\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.837904 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.839867 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4536236-3a5b-4b7c-8150-aa52693e03cb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv\" (UID: \"f4536236-3a5b-4b7c-8150-aa52693e03cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.839948 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b3ecc0f-c11f-4572-8303-65ba8f955b03-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb\" (UID: \"6b3ecc0f-c11f-4572-8303-65ba8f955b03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.840065 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4536236-3a5b-4b7c-8150-aa52693e03cb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv\" (UID: \"f4536236-3a5b-4b7c-8150-aa52693e03cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.840127 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b3ecc0f-c11f-4572-8303-65ba8f955b03-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb\" (UID: \"6b3ecc0f-c11f-4572-8303-65ba8f955b03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.897305 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b3ecc0f-c11f-4572-8303-65ba8f955b03-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb\" (UID: \"6b3ecc0f-c11f-4572-8303-65ba8f955b03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.903621 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b3ecc0f-c11f-4572-8303-65ba8f955b03-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb\" (UID: \"6b3ecc0f-c11f-4572-8303-65ba8f955b03\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.941068 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-5nfbm"] Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.942566 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.944677 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4536236-3a5b-4b7c-8150-aa52693e03cb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv\" (UID: \"f4536236-3a5b-4b7c-8150-aa52693e03cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.944744 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4536236-3a5b-4b7c-8150-aa52693e03cb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv\" (UID: \"f4536236-3a5b-4b7c-8150-aa52693e03cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.948297 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-5nfbm"] Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.951399 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-8zpss" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.951616 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.986460 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.990839 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4536236-3a5b-4b7c-8150-aa52693e03cb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv\" (UID: \"f4536236-3a5b-4b7c-8150-aa52693e03cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" Oct 04 04:50:06 crc kubenswrapper[4770]: I1004 04:50:06.991405 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4536236-3a5b-4b7c-8150-aa52693e03cb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv\" (UID: \"f4536236-3a5b-4b7c-8150-aa52693e03cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.050610 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e942fae-95d7-4a1c-bd99-d049c23d15cd-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-5nfbm\" (UID: \"3e942fae-95d7-4a1c-bd99-d049c23d15cd\") " pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.051026 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-589zh\" (UniqueName: \"kubernetes.io/projected/3e942fae-95d7-4a1c-bd99-d049c23d15cd-kube-api-access-589zh\") pod \"observability-operator-cc5f78dfc-5nfbm\" (UID: \"3e942fae-95d7-4a1c-bd99-d049c23d15cd\") " pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.095115 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-qwcjq"] Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.096528 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.102701 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-bwzbb" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.110546 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-qwcjq"] Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.155375 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-589zh\" (UniqueName: \"kubernetes.io/projected/3e942fae-95d7-4a1c-bd99-d049c23d15cd-kube-api-access-589zh\") pod \"observability-operator-cc5f78dfc-5nfbm\" (UID: \"3e942fae-95d7-4a1c-bd99-d049c23d15cd\") " pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.155457 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e942fae-95d7-4a1c-bd99-d049c23d15cd-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-5nfbm\" (UID: \"3e942fae-95d7-4a1c-bd99-d049c23d15cd\") " pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.165099 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e942fae-95d7-4a1c-bd99-d049c23d15cd-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-5nfbm\" (UID: \"3e942fae-95d7-4a1c-bd99-d049c23d15cd\") " pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.183802 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-589zh\" (UniqueName: \"kubernetes.io/projected/3e942fae-95d7-4a1c-bd99-d049c23d15cd-kube-api-access-589zh\") pod \"observability-operator-cc5f78dfc-5nfbm\" (UID: \"3e942fae-95d7-4a1c-bd99-d049c23d15cd\") " pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.205441 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.256865 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/54e760b6-1a04-4551-8fc4-d461748805d4-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-qwcjq\" (UID: \"54e760b6-1a04-4551-8fc4-d461748805d4\") " pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.256926 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxpbl\" (UniqueName: \"kubernetes.io/projected/54e760b6-1a04-4551-8fc4-d461748805d4-kube-api-access-fxpbl\") pod \"perses-operator-54bc95c9fb-qwcjq\" (UID: \"54e760b6-1a04-4551-8fc4-d461748805d4\") " pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.359295 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/54e760b6-1a04-4551-8fc4-d461748805d4-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-qwcjq\" (UID: \"54e760b6-1a04-4551-8fc4-d461748805d4\") " pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.359604 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxpbl\" (UniqueName: \"kubernetes.io/projected/54e760b6-1a04-4551-8fc4-d461748805d4-kube-api-access-fxpbl\") pod \"perses-operator-54bc95c9fb-qwcjq\" (UID: \"54e760b6-1a04-4551-8fc4-d461748805d4\") " pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.360497 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/54e760b6-1a04-4551-8fc4-d461748805d4-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-qwcjq\" (UID: \"54e760b6-1a04-4551-8fc4-d461748805d4\") " pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.383973 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxpbl\" (UniqueName: \"kubernetes.io/projected/54e760b6-1a04-4551-8fc4-d461748805d4-kube-api-access-fxpbl\") pod \"perses-operator-54bc95c9fb-qwcjq\" (UID: \"54e760b6-1a04-4551-8fc4-d461748805d4\") " pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.391432 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.450329 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.503583 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv"] Oct 04 04:50:07 crc kubenswrapper[4770]: W1004 04:50:07.525159 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb55bd1e1_8afd_42a0_b180_c91c3f66fb4d.slice/crio-68591d1cb7c9483e47a937d903ce0851c9a67c75d4cce7a5a979d8768b1d9714 WatchSource:0}: Error finding container 68591d1cb7c9483e47a937d903ce0851c9a67c75d4cce7a5a979d8768b1d9714: Status 404 returned error can't find the container with id 68591d1cb7c9483e47a937d903ce0851c9a67c75d4cce7a5a979d8768b1d9714 Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.736152 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb"] Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.838284 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv"] Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.955144 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv" event={"ID":"b55bd1e1-8afd-42a0-b180-c91c3f66fb4d","Type":"ContainerStarted","Data":"68591d1cb7c9483e47a937d903ce0851c9a67c75d4cce7a5a979d8768b1d9714"} Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.956541 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" event={"ID":"f4536236-3a5b-4b7c-8150-aa52693e03cb","Type":"ContainerStarted","Data":"4b2235f60ac33311a4856e61838089f17b171919030bfb503d07b5ae43dec3be"} Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.958027 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" event={"ID":"6b3ecc0f-c11f-4572-8303-65ba8f955b03","Type":"ContainerStarted","Data":"2ec2dd4b8d41aea6e6861b8e6de0d5e03bbf10c35352aecef704a732fa10d2a2"} Oct 04 04:50:07 crc kubenswrapper[4770]: I1004 04:50:07.960558 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-5nfbm"] Oct 04 04:50:07 crc kubenswrapper[4770]: W1004 04:50:07.961717 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e942fae_95d7_4a1c_bd99_d049c23d15cd.slice/crio-fdb89e36e13c0f494ade03fd82b71a5d41d1a2f59d71a887cb88e6313afa61b8 WatchSource:0}: Error finding container fdb89e36e13c0f494ade03fd82b71a5d41d1a2f59d71a887cb88e6313afa61b8: Status 404 returned error can't find the container with id fdb89e36e13c0f494ade03fd82b71a5d41d1a2f59d71a887cb88e6313afa61b8 Oct 04 04:50:08 crc kubenswrapper[4770]: I1004 04:50:08.076482 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-qwcjq"] Oct 04 04:50:08 crc kubenswrapper[4770]: W1004 04:50:08.091450 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54e760b6_1a04_4551_8fc4_d461748805d4.slice/crio-52d8fd41cfe2c89aa9e099a9194e0a843483708b5a55d6bc98a52fde3b67723b WatchSource:0}: Error finding container 52d8fd41cfe2c89aa9e099a9194e0a843483708b5a55d6bc98a52fde3b67723b: Status 404 returned error can't find the container with id 52d8fd41cfe2c89aa9e099a9194e0a843483708b5a55d6bc98a52fde3b67723b Oct 04 04:50:08 crc kubenswrapper[4770]: I1004 04:50:08.982561 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" event={"ID":"54e760b6-1a04-4551-8fc4-d461748805d4","Type":"ContainerStarted","Data":"52d8fd41cfe2c89aa9e099a9194e0a843483708b5a55d6bc98a52fde3b67723b"} Oct 04 04:50:08 crc kubenswrapper[4770]: I1004 04:50:08.985377 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" event={"ID":"3e942fae-95d7-4a1c-bd99-d049c23d15cd","Type":"ContainerStarted","Data":"fdb89e36e13c0f494ade03fd82b71a5d41d1a2f59d71a887cb88e6313afa61b8"} Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.107045 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" event={"ID":"3e942fae-95d7-4a1c-bd99-d049c23d15cd","Type":"ContainerStarted","Data":"1e05a1b74b23299af4c1784b437f94ac06dd9967002dab237c61ad9dd6670d71"} Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.107613 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.108819 4770 patch_prober.go:28] interesting pod/observability-operator-cc5f78dfc-5nfbm container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.129:8081/healthz\": dial tcp 10.217.1.129:8081: connect: connection refused" start-of-body= Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.108891 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" podUID="3e942fae-95d7-4a1c-bd99-d049c23d15cd" containerName="operator" probeResult="failure" output="Get \"http://10.217.1.129:8081/healthz\": dial tcp 10.217.1.129:8081: connect: connection refused" Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.109414 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" event={"ID":"f4536236-3a5b-4b7c-8150-aa52693e03cb","Type":"ContainerStarted","Data":"5634ea15e2fe3f5793e9644cd9fedef72d612f90b0e4fed8a17f34e97738eba0"} Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.110891 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" event={"ID":"54e760b6-1a04-4551-8fc4-d461748805d4","Type":"ContainerStarted","Data":"28b319425f5208c3afde8365396e8ea2a4723bcd182efc28d982dcbca531d9f4"} Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.110994 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.112402 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" event={"ID":"6b3ecc0f-c11f-4572-8303-65ba8f955b03","Type":"ContainerStarted","Data":"2fadf2a8ef91fa1e7d99ec3b42bfc5f4df363cfc6554aadb1fa9d9a0253bca9b"} Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.114768 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv" event={"ID":"b55bd1e1-8afd-42a0-b180-c91c3f66fb4d","Type":"ContainerStarted","Data":"64850311fc8f43fa1fb139450e72ef74fc8b4790fbf866e8abe1baafb225837d"} Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.137612 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" podStartSLOduration=2.849701548 podStartE2EDuration="14.137593155s" podCreationTimestamp="2025-10-04 04:50:06 +0000 UTC" firstStartedPulling="2025-10-04 04:50:07.964265889 +0000 UTC m=+6419.256275601" lastFinishedPulling="2025-10-04 04:50:19.252157496 +0000 UTC m=+6430.544167208" observedRunningTime="2025-10-04 04:50:20.130574931 +0000 UTC m=+6431.422584653" watchObservedRunningTime="2025-10-04 04:50:20.137593155 +0000 UTC m=+6431.429602877" Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.156938 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb" podStartSLOduration=2.7031850349999997 podStartE2EDuration="14.156922221s" podCreationTimestamp="2025-10-04 04:50:06 +0000 UTC" firstStartedPulling="2025-10-04 04:50:07.740903096 +0000 UTC m=+6419.032912808" lastFinishedPulling="2025-10-04 04:50:19.194640282 +0000 UTC m=+6430.486649994" observedRunningTime="2025-10-04 04:50:20.154306942 +0000 UTC m=+6431.446316674" watchObservedRunningTime="2025-10-04 04:50:20.156922221 +0000 UTC m=+6431.448931933" Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.191300 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" podStartSLOduration=2.088914314 podStartE2EDuration="13.19128464s" podCreationTimestamp="2025-10-04 04:50:07 +0000 UTC" firstStartedPulling="2025-10-04 04:50:08.093429047 +0000 UTC m=+6419.385438759" lastFinishedPulling="2025-10-04 04:50:19.195799373 +0000 UTC m=+6430.487809085" observedRunningTime="2025-10-04 04:50:20.19013823 +0000 UTC m=+6431.482147942" watchObservedRunningTime="2025-10-04 04:50:20.19128464 +0000 UTC m=+6431.483294352" Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.212968 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-d5cfv" podStartSLOduration=2.544597278 podStartE2EDuration="14.212953666s" podCreationTimestamp="2025-10-04 04:50:06 +0000 UTC" firstStartedPulling="2025-10-04 04:50:07.527388462 +0000 UTC m=+6418.819398174" lastFinishedPulling="2025-10-04 04:50:19.19574485 +0000 UTC m=+6430.487754562" observedRunningTime="2025-10-04 04:50:20.208406098 +0000 UTC m=+6431.500415830" watchObservedRunningTime="2025-10-04 04:50:20.212953666 +0000 UTC m=+6431.504963378" Oct 04 04:50:20 crc kubenswrapper[4770]: I1004 04:50:20.237792 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv" podStartSLOduration=2.887164188 podStartE2EDuration="14.237772296s" podCreationTimestamp="2025-10-04 04:50:06 +0000 UTC" firstStartedPulling="2025-10-04 04:50:07.846845448 +0000 UTC m=+6419.138855160" lastFinishedPulling="2025-10-04 04:50:19.197453546 +0000 UTC m=+6430.489463268" observedRunningTime="2025-10-04 04:50:20.234853769 +0000 UTC m=+6431.526863481" watchObservedRunningTime="2025-10-04 04:50:20.237772296 +0000 UTC m=+6431.529782008" Oct 04 04:50:21 crc kubenswrapper[4770]: I1004 04:50:21.145699 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-5nfbm" Oct 04 04:50:27 crc kubenswrapper[4770]: I1004 04:50:27.453286 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-qwcjq" Oct 04 04:50:28 crc kubenswrapper[4770]: I1004 04:50:28.930796 4770 scope.go:117] "RemoveContainer" containerID="e196645e8452456154ccd1036a1638f33e40f698839d9a335c09babc302dc4a9" Oct 04 04:50:28 crc kubenswrapper[4770]: I1004 04:50:28.966734 4770 scope.go:117] "RemoveContainer" containerID="cf45077d3fae92725c35948ec5d905fd1cc36c0201520ecfc5e506505dc88a89" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.180509 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.180707 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" containerName="openstackclient" containerID="cri-o://bf7f4d11301a1c79acddbb5a17a300c5a528dc95b07e0d485e84fcadfde0a0b9" gracePeriod=2 Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.200962 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.236076 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 04:50:30 crc kubenswrapper[4770]: E1004 04:50:30.236849 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" containerName="openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.236938 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" containerName="openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.237321 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" containerName="openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.238319 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.271959 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.316621 4770 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" podUID="97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.430250 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41-openstack-config-secret\") pod \"openstackclient\" (UID: \"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41\") " pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.430349 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41-openstack-config\") pod \"openstackclient\" (UID: \"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41\") " pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.430783 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4qgc\" (UniqueName: \"kubernetes.io/projected/97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41-kube-api-access-p4qgc\") pod \"openstackclient\" (UID: \"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41\") " pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.449927 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.451173 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.456783 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-rtzfj" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.478082 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.532937 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41-openstack-config\") pod \"openstackclient\" (UID: \"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41\") " pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.533858 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4qgc\" (UniqueName: \"kubernetes.io/projected/97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41-kube-api-access-p4qgc\") pod \"openstackclient\" (UID: \"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41\") " pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.533902 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41-openstack-config-secret\") pod \"openstackclient\" (UID: \"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41\") " pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.534928 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41-openstack-config\") pod \"openstackclient\" (UID: \"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41\") " pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.553487 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41-openstack-config-secret\") pod \"openstackclient\" (UID: \"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41\") " pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.564482 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4qgc\" (UniqueName: \"kubernetes.io/projected/97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41-kube-api-access-p4qgc\") pod \"openstackclient\" (UID: \"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41\") " pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.576464 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.637449 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt9fw\" (UniqueName: \"kubernetes.io/projected/a9e94c14-ec90-4ff2-97a8-8b98ea370af1-kube-api-access-qt9fw\") pod \"kube-state-metrics-0\" (UID: \"a9e94c14-ec90-4ff2-97a8-8b98ea370af1\") " pod="openstack/kube-state-metrics-0" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.738877 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt9fw\" (UniqueName: \"kubernetes.io/projected/a9e94c14-ec90-4ff2-97a8-8b98ea370af1-kube-api-access-qt9fw\") pod \"kube-state-metrics-0\" (UID: \"a9e94c14-ec90-4ff2-97a8-8b98ea370af1\") " pod="openstack/kube-state-metrics-0" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.773750 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt9fw\" (UniqueName: \"kubernetes.io/projected/a9e94c14-ec90-4ff2-97a8-8b98ea370af1-kube-api-access-qt9fw\") pod \"kube-state-metrics-0\" (UID: \"a9e94c14-ec90-4ff2-97a8-8b98ea370af1\") " pod="openstack/kube-state-metrics-0" Oct 04 04:50:30 crc kubenswrapper[4770]: I1004 04:50:30.780747 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.049466 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.066776 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.066893 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.071366 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.071615 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.071818 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-gr2h8" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.079199 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.172490 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/093c0745-f3d5-4954-9617-f69e6b4a5529-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.172568 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/093c0745-f3d5-4954-9617-f69e6b4a5529-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.172653 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/093c0745-f3d5-4954-9617-f69e6b4a5529-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.172676 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/093c0745-f3d5-4954-9617-f69e6b4a5529-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.172724 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/093c0745-f3d5-4954-9617-f69e6b4a5529-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.172761 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr8q7\" (UniqueName: \"kubernetes.io/projected/093c0745-f3d5-4954-9617-f69e6b4a5529-kube-api-access-gr8q7\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.274687 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/093c0745-f3d5-4954-9617-f69e6b4a5529-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.274959 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/093c0745-f3d5-4954-9617-f69e6b4a5529-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.274988 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/093c0745-f3d5-4954-9617-f69e6b4a5529-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.275051 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr8q7\" (UniqueName: \"kubernetes.io/projected/093c0745-f3d5-4954-9617-f69e6b4a5529-kube-api-access-gr8q7\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.275130 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/093c0745-f3d5-4954-9617-f69e6b4a5529-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.275195 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/093c0745-f3d5-4954-9617-f69e6b4a5529-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.277488 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/093c0745-f3d5-4954-9617-f69e6b4a5529-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.281065 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/093c0745-f3d5-4954-9617-f69e6b4a5529-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.282720 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/093c0745-f3d5-4954-9617-f69e6b4a5529-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.285321 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/093c0745-f3d5-4954-9617-f69e6b4a5529-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.286936 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/093c0745-f3d5-4954-9617-f69e6b4a5529-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.317584 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr8q7\" (UniqueName: \"kubernetes.io/projected/093c0745-f3d5-4954-9617-f69e6b4a5529-kube-api-access-gr8q7\") pod \"alertmanager-metric-storage-0\" (UID: \"093c0745-f3d5-4954-9617-f69e6b4a5529\") " pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.413548 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.622664 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.682234 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.773725 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.784125 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.792982 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.793455 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.793700 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.793936 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mx7zg" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.794176 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.805676 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.862089 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.908528 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4zh6\" (UniqueName: \"kubernetes.io/projected/00a56b1a-324e-4d06-a234-0821994a1a48-kube-api-access-t4zh6\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.919369 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/00a56b1a-324e-4d06-a234-0821994a1a48-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.919642 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/00a56b1a-324e-4d06-a234-0821994a1a48-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.919778 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/00a56b1a-324e-4d06-a234-0821994a1a48-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.919910 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/00a56b1a-324e-4d06-a234-0821994a1a48-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.920075 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bc51a319-dea0-45c8-a350-1e5777816c83\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bc51a319-dea0-45c8-a350-1e5777816c83\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.920237 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/00a56b1a-324e-4d06-a234-0821994a1a48-config\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:31 crc kubenswrapper[4770]: I1004 04:50:31.920353 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/00a56b1a-324e-4d06-a234-0821994a1a48-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.025776 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/00a56b1a-324e-4d06-a234-0821994a1a48-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.026736 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4zh6\" (UniqueName: \"kubernetes.io/projected/00a56b1a-324e-4d06-a234-0821994a1a48-kube-api-access-t4zh6\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.026998 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/00a56b1a-324e-4d06-a234-0821994a1a48-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.027117 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/00a56b1a-324e-4d06-a234-0821994a1a48-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.029523 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/00a56b1a-324e-4d06-a234-0821994a1a48-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.029658 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/00a56b1a-324e-4d06-a234-0821994a1a48-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.029820 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bc51a319-dea0-45c8-a350-1e5777816c83\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bc51a319-dea0-45c8-a350-1e5777816c83\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.029975 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/00a56b1a-324e-4d06-a234-0821994a1a48-config\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.031893 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/00a56b1a-324e-4d06-a234-0821994a1a48-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.035112 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/00a56b1a-324e-4d06-a234-0821994a1a48-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.038557 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/00a56b1a-324e-4d06-a234-0821994a1a48-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.038997 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/00a56b1a-324e-4d06-a234-0821994a1a48-config\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.065190 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/00a56b1a-324e-4d06-a234-0821994a1a48-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.066890 4770 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.066950 4770 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bc51a319-dea0-45c8-a350-1e5777816c83\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bc51a319-dea0-45c8-a350-1e5777816c83\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bce595cd9a373fd6397c1ff08893990a909645a7118903153b97994fd38edba9/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.081254 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4zh6\" (UniqueName: \"kubernetes.io/projected/00a56b1a-324e-4d06-a234-0821994a1a48-kube-api-access-t4zh6\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.092369 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/00a56b1a-324e-4d06-a234-0821994a1a48-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.227516 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bc51a319-dea0-45c8-a350-1e5777816c83\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bc51a319-dea0-45c8-a350-1e5777816c83\") pod \"prometheus-metric-storage-0\" (UID: \"00a56b1a-324e-4d06-a234-0821994a1a48\") " pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.295054 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.303148 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a9e94c14-ec90-4ff2-97a8-8b98ea370af1","Type":"ContainerStarted","Data":"68e55298b81b50ccbabbbbe04b268ef797dd52f5957bab471d3b8ad989bb815f"} Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.309263 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41","Type":"ContainerStarted","Data":"b0d048594dad73c172e31cad311fe4094f8ca283ad50cd56be69abc5c93d5827"} Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.315498 4770 generic.go:334] "Generic (PLEG): container finished" podID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" containerID="bf7f4d11301a1c79acddbb5a17a300c5a528dc95b07e0d485e84fcadfde0a0b9" exitCode=137 Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.458361 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.535614 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.656770 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config\") pod \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.656871 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config-secret\") pod \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.657021 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqrxc\" (UniqueName: \"kubernetes.io/projected/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-kube-api-access-kqrxc\") pod \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\" (UID: \"1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7\") " Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.695176 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-kube-api-access-kqrxc" (OuterVolumeSpecName: "kube-api-access-kqrxc") pod "1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" (UID: "1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7"). InnerVolumeSpecName "kube-api-access-kqrxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.714618 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" (UID: "1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.740088 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" (UID: "1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.759930 4770 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.759968 4770 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.759982 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqrxc\" (UniqueName: \"kubernetes.io/projected/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7-kube-api-access-kqrxc\") on node \"crc\" DevicePath \"\"" Oct 04 04:50:32 crc kubenswrapper[4770]: I1004 04:50:32.952435 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.332076 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41","Type":"ContainerStarted","Data":"e9b82eb4dc08d887642d6ac8406d3deb75c080900b3ed5c36f88a89eabe31b2b"} Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.335928 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"00a56b1a-324e-4d06-a234-0821994a1a48","Type":"ContainerStarted","Data":"2229157f924a59cdfc210c111c3fdd041b63930622fb942ab0660c198e71daf4"} Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.337612 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"093c0745-f3d5-4954-9617-f69e6b4a5529","Type":"ContainerStarted","Data":"6e3faee53528cf9d2d5e0db5ac5acdc4408b91b9221517d73c17461e2ec6be10"} Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.339308 4770 scope.go:117] "RemoveContainer" containerID="bf7f4d11301a1c79acddbb5a17a300c5a528dc95b07e0d485e84fcadfde0a0b9" Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.339460 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.350284 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a9e94c14-ec90-4ff2-97a8-8b98ea370af1","Type":"ContainerStarted","Data":"8517fa273fc6baa81694953e6c633b74eaf2a6ccf842ba4fdda08988483da68f"} Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.350442 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.360052 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.359994381 podStartE2EDuration="3.359994381s" podCreationTimestamp="2025-10-04 04:50:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:50:33.350392579 +0000 UTC m=+6444.642402301" watchObservedRunningTime="2025-10-04 04:50:33.359994381 +0000 UTC m=+6444.652004093" Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.371188 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.747015458 podStartE2EDuration="3.371171603s" podCreationTimestamp="2025-10-04 04:50:30 +0000 UTC" firstStartedPulling="2025-10-04 04:50:31.708823383 +0000 UTC m=+6443.000833095" lastFinishedPulling="2025-10-04 04:50:32.332979528 +0000 UTC m=+6443.624989240" observedRunningTime="2025-10-04 04:50:33.368412571 +0000 UTC m=+6444.660422273" watchObservedRunningTime="2025-10-04 04:50:33.371171603 +0000 UTC m=+6444.663181315" Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.371432 4770 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" podUID="97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41" Oct 04 04:50:33 crc kubenswrapper[4770]: I1004 04:50:33.698982 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7" path="/var/lib/kubelet/pods/1a43bcdb-c35a-4dac-9cbe-3188ceca6cc7/volumes" Oct 04 04:50:39 crc kubenswrapper[4770]: I1004 04:50:39.422539 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"00a56b1a-324e-4d06-a234-0821994a1a48","Type":"ContainerStarted","Data":"d0656e07e95db17ef453cbaa434bb2542ed684711093a1acb41554c13e936e59"} Oct 04 04:50:39 crc kubenswrapper[4770]: I1004 04:50:39.428731 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"093c0745-f3d5-4954-9617-f69e6b4a5529","Type":"ContainerStarted","Data":"bf2171a333d180d3dc052a78cb29186756534f383e3778affa9ab4c650068fb3"} Oct 04 04:50:40 crc kubenswrapper[4770]: I1004 04:50:40.785860 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 04:50:45 crc kubenswrapper[4770]: I1004 04:50:45.484596 4770 generic.go:334] "Generic (PLEG): container finished" podID="00a56b1a-324e-4d06-a234-0821994a1a48" containerID="d0656e07e95db17ef453cbaa434bb2542ed684711093a1acb41554c13e936e59" exitCode=0 Oct 04 04:50:45 crc kubenswrapper[4770]: I1004 04:50:45.484688 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"00a56b1a-324e-4d06-a234-0821994a1a48","Type":"ContainerDied","Data":"d0656e07e95db17ef453cbaa434bb2542ed684711093a1acb41554c13e936e59"} Oct 04 04:50:45 crc kubenswrapper[4770]: I1004 04:50:45.486735 4770 generic.go:334] "Generic (PLEG): container finished" podID="093c0745-f3d5-4954-9617-f69e6b4a5529" containerID="bf2171a333d180d3dc052a78cb29186756534f383e3778affa9ab4c650068fb3" exitCode=0 Oct 04 04:50:45 crc kubenswrapper[4770]: I1004 04:50:45.486760 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"093c0745-f3d5-4954-9617-f69e6b4a5529","Type":"ContainerDied","Data":"bf2171a333d180d3dc052a78cb29186756534f383e3778affa9ab4c650068fb3"} Oct 04 04:50:53 crc kubenswrapper[4770]: I1004 04:50:53.576863 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"00a56b1a-324e-4d06-a234-0821994a1a48","Type":"ContainerStarted","Data":"f4738699b565a0bc2f241df393ab813ad9283125f7672fe3b1ff570e2b147fe5"} Oct 04 04:50:53 crc kubenswrapper[4770]: I1004 04:50:53.581321 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"093c0745-f3d5-4954-9617-f69e6b4a5529","Type":"ContainerStarted","Data":"baeb1cdc9011f5d2a4dab1630d88d413851c33328746b92711028e473b8cbeb1"} Oct 04 04:50:57 crc kubenswrapper[4770]: I1004 04:50:57.626718 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"00a56b1a-324e-4d06-a234-0821994a1a48","Type":"ContainerStarted","Data":"eb3ea551643856df2a282658bbc4b0d67a6ca5c67f14579235250df8b25791f8"} Oct 04 04:50:57 crc kubenswrapper[4770]: I1004 04:50:57.633582 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"093c0745-f3d5-4954-9617-f69e6b4a5529","Type":"ContainerStarted","Data":"4f698ebd3981e7ff26ae52178faba5eb45c94bf6a81260e3774fb39bcb670040"} Oct 04 04:50:57 crc kubenswrapper[4770]: I1004 04:50:57.634176 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:57 crc kubenswrapper[4770]: I1004 04:50:57.638938 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 04 04:50:57 crc kubenswrapper[4770]: I1004 04:50:57.667570 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=7.853937232 podStartE2EDuration="27.667554213s" podCreationTimestamp="2025-10-04 04:50:30 +0000 UTC" firstStartedPulling="2025-10-04 04:50:32.487246973 +0000 UTC m=+6443.779256675" lastFinishedPulling="2025-10-04 04:50:52.300863924 +0000 UTC m=+6463.592873656" observedRunningTime="2025-10-04 04:50:57.666305471 +0000 UTC m=+6468.958315183" watchObservedRunningTime="2025-10-04 04:50:57.667554213 +0000 UTC m=+6468.959563925" Oct 04 04:51:05 crc kubenswrapper[4770]: I1004 04:51:05.724315 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"00a56b1a-324e-4d06-a234-0821994a1a48","Type":"ContainerStarted","Data":"24b7e083ef9f0153d455266b06fa79030d1d845120ad3ceb460fa12c6f50974e"} Oct 04 04:51:05 crc kubenswrapper[4770]: I1004 04:51:05.758836 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.819287092 podStartE2EDuration="35.758810661s" podCreationTimestamp="2025-10-04 04:50:30 +0000 UTC" firstStartedPulling="2025-10-04 04:50:32.937458109 +0000 UTC m=+6444.229467821" lastFinishedPulling="2025-10-04 04:51:04.876981678 +0000 UTC m=+6476.168991390" observedRunningTime="2025-10-04 04:51:05.755340101 +0000 UTC m=+6477.047349843" watchObservedRunningTime="2025-10-04 04:51:05.758810661 +0000 UTC m=+6477.050820373" Oct 04 04:51:07 crc kubenswrapper[4770]: I1004 04:51:07.295768 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.651839 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.654258 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.656368 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.657079 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.670065 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.740948 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-log-httpd\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.741130 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.741206 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-scripts\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.741270 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-run-httpd\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.741475 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.741965 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-config-data\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.742092 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr8qs\" (UniqueName: \"kubernetes.io/projected/493e4b1a-4c52-46c8-9a68-b42590334563-kube-api-access-qr8qs\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.844127 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.844228 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-config-data\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.844296 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr8qs\" (UniqueName: \"kubernetes.io/projected/493e4b1a-4c52-46c8-9a68-b42590334563-kube-api-access-qr8qs\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.844402 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-log-httpd\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.844454 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.844481 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-scripts\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.844514 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-run-httpd\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.845058 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-run-httpd\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.845199 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-log-httpd\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.865552 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.865705 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.865955 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-scripts\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.866628 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-config-data\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.887252 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr8qs\" (UniqueName: \"kubernetes.io/projected/493e4b1a-4c52-46c8-9a68-b42590334563-kube-api-access-qr8qs\") pod \"ceilometer-0\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " pod="openstack/ceilometer-0" Oct 04 04:51:08 crc kubenswrapper[4770]: I1004 04:51:08.986208 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:51:09 crc kubenswrapper[4770]: I1004 04:51:09.579771 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:51:09 crc kubenswrapper[4770]: I1004 04:51:09.772453 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerStarted","Data":"9eb0b9eb605ee5a7e187656dd2b961b2255cc88a693848d4b9ae8eccbca26b5e"} Oct 04 04:51:11 crc kubenswrapper[4770]: I1004 04:51:11.790337 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerStarted","Data":"4165e486e97d2038dba6cb027237224a9f1953fa61dfd4c81054758cd3ff172e"} Oct 04 04:51:12 crc kubenswrapper[4770]: I1004 04:51:12.803135 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerStarted","Data":"8733e2b303662d51be26db30365ad2ef85c06839d42ebec6faab30046e2e49f9"} Oct 04 04:51:13 crc kubenswrapper[4770]: I1004 04:51:13.831776 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerStarted","Data":"e972fc747605f87af6d1f46701fc85036d0c08dba48767c4d45a22ce9768f0bc"} Oct 04 04:51:16 crc kubenswrapper[4770]: I1004 04:51:16.867602 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerStarted","Data":"4e0a1355063d4720481cd7a566e3039a17939e0e7b17d266652b91b9f18e8ff8"} Oct 04 04:51:16 crc kubenswrapper[4770]: I1004 04:51:16.868244 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:51:16 crc kubenswrapper[4770]: I1004 04:51:16.891871 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.204827499 podStartE2EDuration="8.891857s" podCreationTimestamp="2025-10-04 04:51:08 +0000 UTC" firstStartedPulling="2025-10-04 04:51:09.59275231 +0000 UTC m=+6480.884762022" lastFinishedPulling="2025-10-04 04:51:16.279781791 +0000 UTC m=+6487.571791523" observedRunningTime="2025-10-04 04:51:16.891675655 +0000 UTC m=+6488.183685377" watchObservedRunningTime="2025-10-04 04:51:16.891857 +0000 UTC m=+6488.183866712" Oct 04 04:51:17 crc kubenswrapper[4770]: I1004 04:51:17.296241 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 04 04:51:17 crc kubenswrapper[4770]: I1004 04:51:17.298378 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 04 04:51:17 crc kubenswrapper[4770]: I1004 04:51:17.880205 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 04 04:51:19 crc kubenswrapper[4770]: I1004 04:51:19.060310 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-4t2xt"] Oct 04 04:51:19 crc kubenswrapper[4770]: I1004 04:51:19.062042 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4t2xt" Oct 04 04:51:19 crc kubenswrapper[4770]: I1004 04:51:19.070566 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-4t2xt"] Oct 04 04:51:19 crc kubenswrapper[4770]: I1004 04:51:19.111499 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nlhq\" (UniqueName: \"kubernetes.io/projected/1ca7fd2c-4805-497e-bbdf-fecba44d2a76-kube-api-access-9nlhq\") pod \"aodh-db-create-4t2xt\" (UID: \"1ca7fd2c-4805-497e-bbdf-fecba44d2a76\") " pod="openstack/aodh-db-create-4t2xt" Oct 04 04:51:19 crc kubenswrapper[4770]: I1004 04:51:19.213643 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nlhq\" (UniqueName: \"kubernetes.io/projected/1ca7fd2c-4805-497e-bbdf-fecba44d2a76-kube-api-access-9nlhq\") pod \"aodh-db-create-4t2xt\" (UID: \"1ca7fd2c-4805-497e-bbdf-fecba44d2a76\") " pod="openstack/aodh-db-create-4t2xt" Oct 04 04:51:19 crc kubenswrapper[4770]: I1004 04:51:19.232556 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nlhq\" (UniqueName: \"kubernetes.io/projected/1ca7fd2c-4805-497e-bbdf-fecba44d2a76-kube-api-access-9nlhq\") pod \"aodh-db-create-4t2xt\" (UID: \"1ca7fd2c-4805-497e-bbdf-fecba44d2a76\") " pod="openstack/aodh-db-create-4t2xt" Oct 04 04:51:19 crc kubenswrapper[4770]: I1004 04:51:19.395236 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4t2xt" Oct 04 04:51:20 crc kubenswrapper[4770]: I1004 04:51:20.054693 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-4t2xt"] Oct 04 04:51:20 crc kubenswrapper[4770]: W1004 04:51:20.065193 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ca7fd2c_4805_497e_bbdf_fecba44d2a76.slice/crio-4341414f620ca1a469e96c11a2d99b5c1a8e329445016922a4d0998152c3fe3c WatchSource:0}: Error finding container 4341414f620ca1a469e96c11a2d99b5c1a8e329445016922a4d0998152c3fe3c: Status 404 returned error can't find the container with id 4341414f620ca1a469e96c11a2d99b5c1a8e329445016922a4d0998152c3fe3c Oct 04 04:51:20 crc kubenswrapper[4770]: I1004 04:51:20.907517 4770 generic.go:334] "Generic (PLEG): container finished" podID="1ca7fd2c-4805-497e-bbdf-fecba44d2a76" containerID="4af590e786535fc2185c53502157d9d8d98b8c4728abb2de52d7a300a0c7d0b2" exitCode=0 Oct 04 04:51:20 crc kubenswrapper[4770]: I1004 04:51:20.907614 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4t2xt" event={"ID":"1ca7fd2c-4805-497e-bbdf-fecba44d2a76","Type":"ContainerDied","Data":"4af590e786535fc2185c53502157d9d8d98b8c4728abb2de52d7a300a0c7d0b2"} Oct 04 04:51:20 crc kubenswrapper[4770]: I1004 04:51:20.907918 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4t2xt" event={"ID":"1ca7fd2c-4805-497e-bbdf-fecba44d2a76","Type":"ContainerStarted","Data":"4341414f620ca1a469e96c11a2d99b5c1a8e329445016922a4d0998152c3fe3c"} Oct 04 04:51:22 crc kubenswrapper[4770]: I1004 04:51:22.337260 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4t2xt" Oct 04 04:51:22 crc kubenswrapper[4770]: I1004 04:51:22.399309 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nlhq\" (UniqueName: \"kubernetes.io/projected/1ca7fd2c-4805-497e-bbdf-fecba44d2a76-kube-api-access-9nlhq\") pod \"1ca7fd2c-4805-497e-bbdf-fecba44d2a76\" (UID: \"1ca7fd2c-4805-497e-bbdf-fecba44d2a76\") " Oct 04 04:51:22 crc kubenswrapper[4770]: I1004 04:51:22.405153 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca7fd2c-4805-497e-bbdf-fecba44d2a76-kube-api-access-9nlhq" (OuterVolumeSpecName: "kube-api-access-9nlhq") pod "1ca7fd2c-4805-497e-bbdf-fecba44d2a76" (UID: "1ca7fd2c-4805-497e-bbdf-fecba44d2a76"). InnerVolumeSpecName "kube-api-access-9nlhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:22 crc kubenswrapper[4770]: I1004 04:51:22.501729 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nlhq\" (UniqueName: \"kubernetes.io/projected/1ca7fd2c-4805-497e-bbdf-fecba44d2a76-kube-api-access-9nlhq\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:22 crc kubenswrapper[4770]: I1004 04:51:22.930022 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4t2xt" event={"ID":"1ca7fd2c-4805-497e-bbdf-fecba44d2a76","Type":"ContainerDied","Data":"4341414f620ca1a469e96c11a2d99b5c1a8e329445016922a4d0998152c3fe3c"} Oct 04 04:51:22 crc kubenswrapper[4770]: I1004 04:51:22.930389 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4341414f620ca1a469e96c11a2d99b5c1a8e329445016922a4d0998152c3fe3c" Oct 04 04:51:22 crc kubenswrapper[4770]: I1004 04:51:22.930090 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4t2xt" Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.298564 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-dcf1-account-create-zzpdm"] Oct 04 04:51:29 crc kubenswrapper[4770]: E1004 04:51:29.299387 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca7fd2c-4805-497e-bbdf-fecba44d2a76" containerName="mariadb-database-create" Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.299398 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca7fd2c-4805-497e-bbdf-fecba44d2a76" containerName="mariadb-database-create" Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.299613 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca7fd2c-4805-497e-bbdf-fecba44d2a76" containerName="mariadb-database-create" Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.300286 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-dcf1-account-create-zzpdm" Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.302119 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.316495 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-dcf1-account-create-zzpdm"] Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.352270 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8fql\" (UniqueName: \"kubernetes.io/projected/7759865b-1f02-4d89-b903-9e49cf7c96ee-kube-api-access-g8fql\") pod \"aodh-dcf1-account-create-zzpdm\" (UID: \"7759865b-1f02-4d89-b903-9e49cf7c96ee\") " pod="openstack/aodh-dcf1-account-create-zzpdm" Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.455233 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8fql\" (UniqueName: \"kubernetes.io/projected/7759865b-1f02-4d89-b903-9e49cf7c96ee-kube-api-access-g8fql\") pod \"aodh-dcf1-account-create-zzpdm\" (UID: \"7759865b-1f02-4d89-b903-9e49cf7c96ee\") " pod="openstack/aodh-dcf1-account-create-zzpdm" Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.476364 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8fql\" (UniqueName: \"kubernetes.io/projected/7759865b-1f02-4d89-b903-9e49cf7c96ee-kube-api-access-g8fql\") pod \"aodh-dcf1-account-create-zzpdm\" (UID: \"7759865b-1f02-4d89-b903-9e49cf7c96ee\") " pod="openstack/aodh-dcf1-account-create-zzpdm" Oct 04 04:51:29 crc kubenswrapper[4770]: I1004 04:51:29.623111 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-dcf1-account-create-zzpdm" Oct 04 04:51:30 crc kubenswrapper[4770]: I1004 04:51:30.191116 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-dcf1-account-create-zzpdm"] Oct 04 04:51:30 crc kubenswrapper[4770]: W1004 04:51:30.201089 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7759865b_1f02_4d89_b903_9e49cf7c96ee.slice/crio-12c919c77e4a6bc4c238b6a6f5f97b3601bc6706d05052c42aecd0929e140022 WatchSource:0}: Error finding container 12c919c77e4a6bc4c238b6a6f5f97b3601bc6706d05052c42aecd0929e140022: Status 404 returned error can't find the container with id 12c919c77e4a6bc4c238b6a6f5f97b3601bc6706d05052c42aecd0929e140022 Oct 04 04:51:31 crc kubenswrapper[4770]: I1004 04:51:31.014119 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-dcf1-account-create-zzpdm" event={"ID":"7759865b-1f02-4d89-b903-9e49cf7c96ee","Type":"ContainerStarted","Data":"6e7bb09c62432c4846dab60223583c09739c12e22a67c5ff62f76ac504d853a2"} Oct 04 04:51:31 crc kubenswrapper[4770]: I1004 04:51:31.014175 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-dcf1-account-create-zzpdm" event={"ID":"7759865b-1f02-4d89-b903-9e49cf7c96ee","Type":"ContainerStarted","Data":"12c919c77e4a6bc4c238b6a6f5f97b3601bc6706d05052c42aecd0929e140022"} Oct 04 04:51:31 crc kubenswrapper[4770]: I1004 04:51:31.034083 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-dcf1-account-create-zzpdm" podStartSLOduration=2.034060313 podStartE2EDuration="2.034060313s" podCreationTimestamp="2025-10-04 04:51:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:51:31.031661661 +0000 UTC m=+6502.323671373" watchObservedRunningTime="2025-10-04 04:51:31.034060313 +0000 UTC m=+6502.326070035" Oct 04 04:51:32 crc kubenswrapper[4770]: I1004 04:51:32.027860 4770 generic.go:334] "Generic (PLEG): container finished" podID="7759865b-1f02-4d89-b903-9e49cf7c96ee" containerID="6e7bb09c62432c4846dab60223583c09739c12e22a67c5ff62f76ac504d853a2" exitCode=0 Oct 04 04:51:32 crc kubenswrapper[4770]: I1004 04:51:32.027955 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-dcf1-account-create-zzpdm" event={"ID":"7759865b-1f02-4d89-b903-9e49cf7c96ee","Type":"ContainerDied","Data":"6e7bb09c62432c4846dab60223583c09739c12e22a67c5ff62f76ac504d853a2"} Oct 04 04:51:33 crc kubenswrapper[4770]: I1004 04:51:33.488679 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-dcf1-account-create-zzpdm" Oct 04 04:51:33 crc kubenswrapper[4770]: I1004 04:51:33.549998 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8fql\" (UniqueName: \"kubernetes.io/projected/7759865b-1f02-4d89-b903-9e49cf7c96ee-kube-api-access-g8fql\") pod \"7759865b-1f02-4d89-b903-9e49cf7c96ee\" (UID: \"7759865b-1f02-4d89-b903-9e49cf7c96ee\") " Oct 04 04:51:33 crc kubenswrapper[4770]: I1004 04:51:33.556518 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7759865b-1f02-4d89-b903-9e49cf7c96ee-kube-api-access-g8fql" (OuterVolumeSpecName: "kube-api-access-g8fql") pod "7759865b-1f02-4d89-b903-9e49cf7c96ee" (UID: "7759865b-1f02-4d89-b903-9e49cf7c96ee"). InnerVolumeSpecName "kube-api-access-g8fql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:33 crc kubenswrapper[4770]: I1004 04:51:33.652656 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8fql\" (UniqueName: \"kubernetes.io/projected/7759865b-1f02-4d89-b903-9e49cf7c96ee-kube-api-access-g8fql\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.048398 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-dcf1-account-create-zzpdm" event={"ID":"7759865b-1f02-4d89-b903-9e49cf7c96ee","Type":"ContainerDied","Data":"12c919c77e4a6bc4c238b6a6f5f97b3601bc6706d05052c42aecd0929e140022"} Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.048449 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12c919c77e4a6bc4c238b6a6f5f97b3601bc6706d05052c42aecd0929e140022" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.048477 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-dcf1-account-create-zzpdm" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.777541 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-tmj47"] Oct 04 04:51:34 crc kubenswrapper[4770]: E1004 04:51:34.778392 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7759865b-1f02-4d89-b903-9e49cf7c96ee" containerName="mariadb-account-create" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.778407 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7759865b-1f02-4d89-b903-9e49cf7c96ee" containerName="mariadb-account-create" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.778641 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7759865b-1f02-4d89-b903-9e49cf7c96ee" containerName="mariadb-account-create" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.779539 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.786597 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.786827 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.787072 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-s6fwf" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.789144 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-tmj47"] Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.882613 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-scripts\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.883058 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-combined-ca-bundle\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.883168 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-config-data\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.883482 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9g7z\" (UniqueName: \"kubernetes.io/projected/c3875920-d1d5-455b-b27e-49599ac8bba9-kube-api-access-q9g7z\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.984937 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-config-data\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.985250 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9g7z\" (UniqueName: \"kubernetes.io/projected/c3875920-d1d5-455b-b27e-49599ac8bba9-kube-api-access-q9g7z\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.985423 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-scripts\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.985510 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-combined-ca-bundle\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.994587 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-combined-ca-bundle\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:34 crc kubenswrapper[4770]: I1004 04:51:34.996059 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-config-data\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:35 crc kubenswrapper[4770]: I1004 04:51:35.006717 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-scripts\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:35 crc kubenswrapper[4770]: I1004 04:51:35.023623 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9g7z\" (UniqueName: \"kubernetes.io/projected/c3875920-d1d5-455b-b27e-49599ac8bba9-kube-api-access-q9g7z\") pod \"aodh-db-sync-tmj47\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:35 crc kubenswrapper[4770]: I1004 04:51:35.104691 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:35 crc kubenswrapper[4770]: I1004 04:51:35.724448 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-tmj47"] Oct 04 04:51:36 crc kubenswrapper[4770]: I1004 04:51:36.078360 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tmj47" event={"ID":"c3875920-d1d5-455b-b27e-49599ac8bba9","Type":"ContainerStarted","Data":"8230bd062cc2a89450d69ecf1690861961907adedbd82c85013f950885889373"} Oct 04 04:51:39 crc kubenswrapper[4770]: I1004 04:51:39.210723 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 04:51:44 crc kubenswrapper[4770]: I1004 04:51:44.208536 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tmj47" event={"ID":"c3875920-d1d5-455b-b27e-49599ac8bba9","Type":"ContainerStarted","Data":"a76bd61b3e309b80e7aa0132849b648cddadc00565f44e1ebab95bdfd2837864"} Oct 04 04:51:44 crc kubenswrapper[4770]: I1004 04:51:44.236150 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-tmj47" podStartSLOduration=3.079832372 podStartE2EDuration="10.236124997s" podCreationTimestamp="2025-10-04 04:51:34 +0000 UTC" firstStartedPulling="2025-10-04 04:51:35.731531647 +0000 UTC m=+6507.023541359" lastFinishedPulling="2025-10-04 04:51:42.887824262 +0000 UTC m=+6514.179833984" observedRunningTime="2025-10-04 04:51:44.225644093 +0000 UTC m=+6515.517653815" watchObservedRunningTime="2025-10-04 04:51:44.236124997 +0000 UTC m=+6515.528134729" Oct 04 04:51:48 crc kubenswrapper[4770]: I1004 04:51:48.252586 4770 generic.go:334] "Generic (PLEG): container finished" podID="c3875920-d1d5-455b-b27e-49599ac8bba9" containerID="a76bd61b3e309b80e7aa0132849b648cddadc00565f44e1ebab95bdfd2837864" exitCode=0 Oct 04 04:51:48 crc kubenswrapper[4770]: I1004 04:51:48.252698 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tmj47" event={"ID":"c3875920-d1d5-455b-b27e-49599ac8bba9","Type":"ContainerDied","Data":"a76bd61b3e309b80e7aa0132849b648cddadc00565f44e1ebab95bdfd2837864"} Oct 04 04:51:49 crc kubenswrapper[4770]: I1004 04:51:49.752786 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:49 crc kubenswrapper[4770]: I1004 04:51:49.928183 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-config-data\") pod \"c3875920-d1d5-455b-b27e-49599ac8bba9\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " Oct 04 04:51:49 crc kubenswrapper[4770]: I1004 04:51:49.929316 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9g7z\" (UniqueName: \"kubernetes.io/projected/c3875920-d1d5-455b-b27e-49599ac8bba9-kube-api-access-q9g7z\") pod \"c3875920-d1d5-455b-b27e-49599ac8bba9\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " Oct 04 04:51:49 crc kubenswrapper[4770]: I1004 04:51:49.929449 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-combined-ca-bundle\") pod \"c3875920-d1d5-455b-b27e-49599ac8bba9\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " Oct 04 04:51:49 crc kubenswrapper[4770]: I1004 04:51:49.929485 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-scripts\") pod \"c3875920-d1d5-455b-b27e-49599ac8bba9\" (UID: \"c3875920-d1d5-455b-b27e-49599ac8bba9\") " Oct 04 04:51:49 crc kubenswrapper[4770]: I1004 04:51:49.934557 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-scripts" (OuterVolumeSpecName: "scripts") pod "c3875920-d1d5-455b-b27e-49599ac8bba9" (UID: "c3875920-d1d5-455b-b27e-49599ac8bba9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:51:49 crc kubenswrapper[4770]: I1004 04:51:49.936301 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3875920-d1d5-455b-b27e-49599ac8bba9-kube-api-access-q9g7z" (OuterVolumeSpecName: "kube-api-access-q9g7z") pod "c3875920-d1d5-455b-b27e-49599ac8bba9" (UID: "c3875920-d1d5-455b-b27e-49599ac8bba9"). InnerVolumeSpecName "kube-api-access-q9g7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:49 crc kubenswrapper[4770]: I1004 04:51:49.960488 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3875920-d1d5-455b-b27e-49599ac8bba9" (UID: "c3875920-d1d5-455b-b27e-49599ac8bba9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:51:49 crc kubenswrapper[4770]: I1004 04:51:49.960984 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-config-data" (OuterVolumeSpecName: "config-data") pod "c3875920-d1d5-455b-b27e-49599ac8bba9" (UID: "c3875920-d1d5-455b-b27e-49599ac8bba9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:51:50 crc kubenswrapper[4770]: I1004 04:51:50.032385 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:50 crc kubenswrapper[4770]: I1004 04:51:50.032737 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:50 crc kubenswrapper[4770]: I1004 04:51:50.032750 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3875920-d1d5-455b-b27e-49599ac8bba9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:50 crc kubenswrapper[4770]: I1004 04:51:50.032759 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9g7z\" (UniqueName: \"kubernetes.io/projected/c3875920-d1d5-455b-b27e-49599ac8bba9-kube-api-access-q9g7z\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:50 crc kubenswrapper[4770]: I1004 04:51:50.274795 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tmj47" event={"ID":"c3875920-d1d5-455b-b27e-49599ac8bba9","Type":"ContainerDied","Data":"8230bd062cc2a89450d69ecf1690861961907adedbd82c85013f950885889373"} Oct 04 04:51:50 crc kubenswrapper[4770]: I1004 04:51:50.274843 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8230bd062cc2a89450d69ecf1690861961907adedbd82c85013f950885889373" Oct 04 04:51:50 crc kubenswrapper[4770]: I1004 04:51:50.274889 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tmj47" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.922927 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 04 04:51:54 crc kubenswrapper[4770]: E1004 04:51:54.925308 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3875920-d1d5-455b-b27e-49599ac8bba9" containerName="aodh-db-sync" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.925416 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3875920-d1d5-455b-b27e-49599ac8bba9" containerName="aodh-db-sync" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.925723 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3875920-d1d5-455b-b27e-49599ac8bba9" containerName="aodh-db-sync" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.927940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.933805 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.934342 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.934494 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-s6fwf" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.937135 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.937185 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-scripts\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.937219 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7wrq\" (UniqueName: \"kubernetes.io/projected/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-kube-api-access-d7wrq\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.937272 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-config-data\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:54 crc kubenswrapper[4770]: I1004 04:51:54.940615 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.039740 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.039956 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-scripts\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.039990 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7wrq\" (UniqueName: \"kubernetes.io/projected/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-kube-api-access-d7wrq\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.040046 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-config-data\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.046969 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-scripts\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.061033 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-config-data\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.061920 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7wrq\" (UniqueName: \"kubernetes.io/projected/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-kube-api-access-d7wrq\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.064990 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f854c7c-1a25-48b9-9bdc-78b4d7ff550e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e\") " pod="openstack/aodh-0" Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.262111 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 04 04:51:55 crc kubenswrapper[4770]: I1004 04:51:55.815392 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 04 04:51:56 crc kubenswrapper[4770]: I1004 04:51:56.330142 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e","Type":"ContainerStarted","Data":"494cf307a70f08e7b673a4a7f5eb94dc4ceeb25674636221613e894a8de0b82f"} Oct 04 04:51:57 crc kubenswrapper[4770]: I1004 04:51:57.952637 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:51:57 crc kubenswrapper[4770]: I1004 04:51:57.953350 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="ceilometer-central-agent" containerID="cri-o://4165e486e97d2038dba6cb027237224a9f1953fa61dfd4c81054758cd3ff172e" gracePeriod=30 Oct 04 04:51:57 crc kubenswrapper[4770]: I1004 04:51:57.953500 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="proxy-httpd" containerID="cri-o://4e0a1355063d4720481cd7a566e3039a17939e0e7b17d266652b91b9f18e8ff8" gracePeriod=30 Oct 04 04:51:57 crc kubenswrapper[4770]: I1004 04:51:57.953547 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="sg-core" containerID="cri-o://e972fc747605f87af6d1f46701fc85036d0c08dba48767c4d45a22ce9768f0bc" gracePeriod=30 Oct 04 04:51:57 crc kubenswrapper[4770]: I1004 04:51:57.953588 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="ceilometer-notification-agent" containerID="cri-o://8733e2b303662d51be26db30365ad2ef85c06839d42ebec6faab30046e2e49f9" gracePeriod=30 Oct 04 04:51:58 crc kubenswrapper[4770]: I1004 04:51:58.355462 4770 generic.go:334] "Generic (PLEG): container finished" podID="493e4b1a-4c52-46c8-9a68-b42590334563" containerID="4e0a1355063d4720481cd7a566e3039a17939e0e7b17d266652b91b9f18e8ff8" exitCode=0 Oct 04 04:51:58 crc kubenswrapper[4770]: I1004 04:51:58.355846 4770 generic.go:334] "Generic (PLEG): container finished" podID="493e4b1a-4c52-46c8-9a68-b42590334563" containerID="e972fc747605f87af6d1f46701fc85036d0c08dba48767c4d45a22ce9768f0bc" exitCode=2 Oct 04 04:51:58 crc kubenswrapper[4770]: I1004 04:51:58.355905 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerDied","Data":"4e0a1355063d4720481cd7a566e3039a17939e0e7b17d266652b91b9f18e8ff8"} Oct 04 04:51:58 crc kubenswrapper[4770]: I1004 04:51:58.355935 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerDied","Data":"e972fc747605f87af6d1f46701fc85036d0c08dba48767c4d45a22ce9768f0bc"} Oct 04 04:51:58 crc kubenswrapper[4770]: I1004 04:51:58.357566 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e","Type":"ContainerStarted","Data":"65a07170d8d53d3c0575f71bbe088b8206737a9becfdce99c0c54a5a27554cf5"} Oct 04 04:51:59 crc kubenswrapper[4770]: I1004 04:51:59.370559 4770 generic.go:334] "Generic (PLEG): container finished" podID="493e4b1a-4c52-46c8-9a68-b42590334563" containerID="4165e486e97d2038dba6cb027237224a9f1953fa61dfd4c81054758cd3ff172e" exitCode=0 Oct 04 04:51:59 crc kubenswrapper[4770]: I1004 04:51:59.370630 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerDied","Data":"4165e486e97d2038dba6cb027237224a9f1953fa61dfd4c81054758cd3ff172e"} Oct 04 04:52:01 crc kubenswrapper[4770]: I1004 04:52:01.796063 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:52:01 crc kubenswrapper[4770]: I1004 04:52:01.796364 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:52:02 crc kubenswrapper[4770]: I1004 04:52:02.408381 4770 generic.go:334] "Generic (PLEG): container finished" podID="493e4b1a-4c52-46c8-9a68-b42590334563" containerID="8733e2b303662d51be26db30365ad2ef85c06839d42ebec6faab30046e2e49f9" exitCode=0 Oct 04 04:52:02 crc kubenswrapper[4770]: I1004 04:52:02.408605 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerDied","Data":"8733e2b303662d51be26db30365ad2ef85c06839d42ebec6faab30046e2e49f9"} Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.095712 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.191654 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-run-httpd\") pod \"493e4b1a-4c52-46c8-9a68-b42590334563\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.192091 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-config-data\") pod \"493e4b1a-4c52-46c8-9a68-b42590334563\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.192470 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-sg-core-conf-yaml\") pod \"493e4b1a-4c52-46c8-9a68-b42590334563\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.192479 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "493e4b1a-4c52-46c8-9a68-b42590334563" (UID: "493e4b1a-4c52-46c8-9a68-b42590334563"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.192965 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-combined-ca-bundle\") pod \"493e4b1a-4c52-46c8-9a68-b42590334563\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.193401 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-scripts\") pod \"493e4b1a-4c52-46c8-9a68-b42590334563\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.193662 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-log-httpd\") pod \"493e4b1a-4c52-46c8-9a68-b42590334563\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.193871 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr8qs\" (UniqueName: \"kubernetes.io/projected/493e4b1a-4c52-46c8-9a68-b42590334563-kube-api-access-qr8qs\") pod \"493e4b1a-4c52-46c8-9a68-b42590334563\" (UID: \"493e4b1a-4c52-46c8-9a68-b42590334563\") " Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.195127 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "493e4b1a-4c52-46c8-9a68-b42590334563" (UID: "493e4b1a-4c52-46c8-9a68-b42590334563"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.196678 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.196767 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/493e4b1a-4c52-46c8-9a68-b42590334563-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.206346 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-scripts" (OuterVolumeSpecName: "scripts") pod "493e4b1a-4c52-46c8-9a68-b42590334563" (UID: "493e4b1a-4c52-46c8-9a68-b42590334563"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.206628 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493e4b1a-4c52-46c8-9a68-b42590334563-kube-api-access-qr8qs" (OuterVolumeSpecName: "kube-api-access-qr8qs") pod "493e4b1a-4c52-46c8-9a68-b42590334563" (UID: "493e4b1a-4c52-46c8-9a68-b42590334563"). InnerVolumeSpecName "kube-api-access-qr8qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.248441 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "493e4b1a-4c52-46c8-9a68-b42590334563" (UID: "493e4b1a-4c52-46c8-9a68-b42590334563"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.286551 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "493e4b1a-4c52-46c8-9a68-b42590334563" (UID: "493e4b1a-4c52-46c8-9a68-b42590334563"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.298201 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.298229 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.298241 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr8qs\" (UniqueName: \"kubernetes.io/projected/493e4b1a-4c52-46c8-9a68-b42590334563-kube-api-access-qr8qs\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.298254 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.321416 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-config-data" (OuterVolumeSpecName: "config-data") pod "493e4b1a-4c52-46c8-9a68-b42590334563" (UID: "493e4b1a-4c52-46c8-9a68-b42590334563"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.399828 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493e4b1a-4c52-46c8-9a68-b42590334563-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.474065 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"493e4b1a-4c52-46c8-9a68-b42590334563","Type":"ContainerDied","Data":"9eb0b9eb605ee5a7e187656dd2b961b2255cc88a693848d4b9ae8eccbca26b5e"} Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.474147 4770 scope.go:117] "RemoveContainer" containerID="4e0a1355063d4720481cd7a566e3039a17939e0e7b17d266652b91b9f18e8ff8" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.474143 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.511054 4770 scope.go:117] "RemoveContainer" containerID="e972fc747605f87af6d1f46701fc85036d0c08dba48767c4d45a22ce9768f0bc" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.554120 4770 scope.go:117] "RemoveContainer" containerID="8733e2b303662d51be26db30365ad2ef85c06839d42ebec6faab30046e2e49f9" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.575864 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.589002 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.614372 4770 scope.go:117] "RemoveContainer" containerID="4165e486e97d2038dba6cb027237224a9f1953fa61dfd4c81054758cd3ff172e" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.619618 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:52:07 crc kubenswrapper[4770]: E1004 04:52:07.620101 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="sg-core" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.620119 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="sg-core" Oct 04 04:52:07 crc kubenswrapper[4770]: E1004 04:52:07.620132 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="ceilometer-notification-agent" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.620139 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="ceilometer-notification-agent" Oct 04 04:52:07 crc kubenswrapper[4770]: E1004 04:52:07.620172 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="ceilometer-central-agent" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.620179 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="ceilometer-central-agent" Oct 04 04:52:07 crc kubenswrapper[4770]: E1004 04:52:07.620195 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="proxy-httpd" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.620203 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="proxy-httpd" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.620397 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="ceilometer-central-agent" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.620425 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="sg-core" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.620436 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="proxy-httpd" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.620447 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" containerName="ceilometer-notification-agent" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.622604 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.624777 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.625723 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.630321 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.684701 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="493e4b1a-4c52-46c8-9a68-b42590334563" path="/var/lib/kubelet/pods/493e4b1a-4c52-46c8-9a68-b42590334563/volumes" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.717221 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-run-httpd\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.717423 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.718846 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s2c2\" (UniqueName: \"kubernetes.io/projected/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-kube-api-access-2s2c2\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.721090 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.721486 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-log-httpd\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.721545 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-scripts\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.722370 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-config-data\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: E1004 04:52:07.775392 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod493e4b1a_4c52_46c8_9a68_b42590334563.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod493e4b1a_4c52_46c8_9a68_b42590334563.slice/crio-9eb0b9eb605ee5a7e187656dd2b961b2255cc88a693848d4b9ae8eccbca26b5e\": RecentStats: unable to find data in memory cache]" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.824031 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.824106 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s2c2\" (UniqueName: \"kubernetes.io/projected/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-kube-api-access-2s2c2\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.824174 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.824205 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-log-httpd\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.824221 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-scripts\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.824247 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-config-data\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.824320 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-run-httpd\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.824753 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-run-httpd\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.825334 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-log-httpd\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.831695 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-scripts\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.831705 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.831762 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-config-data\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.831874 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.843982 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s2c2\" (UniqueName: \"kubernetes.io/projected/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-kube-api-access-2s2c2\") pod \"ceilometer-0\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " pod="openstack/ceilometer-0" Oct 04 04:52:07 crc kubenswrapper[4770]: I1004 04:52:07.946756 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:52:08 crc kubenswrapper[4770]: I1004 04:52:08.457862 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:52:08 crc kubenswrapper[4770]: W1004 04:52:08.460027 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc57c72ce_d6a9_45fd_a0e6_ccd7b974380f.slice/crio-a048cafc866b13a2262cc6bb0a9cc80a314a4fc895fd81b7a7812eac720684f1 WatchSource:0}: Error finding container a048cafc866b13a2262cc6bb0a9cc80a314a4fc895fd81b7a7812eac720684f1: Status 404 returned error can't find the container with id a048cafc866b13a2262cc6bb0a9cc80a314a4fc895fd81b7a7812eac720684f1 Oct 04 04:52:08 crc kubenswrapper[4770]: I1004 04:52:08.485417 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e","Type":"ContainerStarted","Data":"f422c366ab147dd5a59eb45b5f094ce9426a3907987b6ba1d59e0de4ccc76f4e"} Oct 04 04:52:08 crc kubenswrapper[4770]: I1004 04:52:08.486573 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerStarted","Data":"a048cafc866b13a2262cc6bb0a9cc80a314a4fc895fd81b7a7812eac720684f1"} Oct 04 04:52:11 crc kubenswrapper[4770]: I1004 04:52:11.534282 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerStarted","Data":"a1079219452f4179d5885f881c017c81f1500a95066d83c2484944ee35e48936"} Oct 04 04:52:12 crc kubenswrapper[4770]: I1004 04:52:12.545397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e","Type":"ContainerStarted","Data":"932fb1338baf2816313372649a806a4062c2d8731e1051b26e590918b45ccd1f"} Oct 04 04:52:15 crc kubenswrapper[4770]: I1004 04:52:15.576706 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerStarted","Data":"1f00cb8a22cc527ca99b4475e123754e10b32963dcc5c396be7763e9230a4a78"} Oct 04 04:52:19 crc kubenswrapper[4770]: I1004 04:52:19.625395 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerStarted","Data":"51894a8e3ad90a2fdd772ce71ebc448c05a9fc853d32f2fbf169b5f4bb01ec54"} Oct 04 04:52:19 crc kubenswrapper[4770]: I1004 04:52:19.627575 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6f854c7c-1a25-48b9-9bdc-78b4d7ff550e","Type":"ContainerStarted","Data":"b7d6266811d136dcb13a2222429facaa0b55f44593f45452a70ae3bc419ff78e"} Oct 04 04:52:19 crc kubenswrapper[4770]: I1004 04:52:19.652661 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.793112288 podStartE2EDuration="25.652645177s" podCreationTimestamp="2025-10-04 04:51:54 +0000 UTC" firstStartedPulling="2025-10-04 04:51:55.820317685 +0000 UTC m=+6527.112327407" lastFinishedPulling="2025-10-04 04:52:18.679850584 +0000 UTC m=+6549.971860296" observedRunningTime="2025-10-04 04:52:19.64705534 +0000 UTC m=+6550.939065052" watchObservedRunningTime="2025-10-04 04:52:19.652645177 +0000 UTC m=+6550.944654889" Oct 04 04:52:22 crc kubenswrapper[4770]: I1004 04:52:22.668079 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerStarted","Data":"b5b3c8a655fe4f33994b15d5d0f43f87d9564556eb7e01563c59a9cf9c10c247"} Oct 04 04:52:22 crc kubenswrapper[4770]: I1004 04:52:22.668784 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:52:22 crc kubenswrapper[4770]: I1004 04:52:22.715677 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.856778998 podStartE2EDuration="15.715656521s" podCreationTimestamp="2025-10-04 04:52:07 +0000 UTC" firstStartedPulling="2025-10-04 04:52:08.46231194 +0000 UTC m=+6539.754321652" lastFinishedPulling="2025-10-04 04:52:22.321189453 +0000 UTC m=+6553.613199175" observedRunningTime="2025-10-04 04:52:22.696818629 +0000 UTC m=+6553.988828341" watchObservedRunningTime="2025-10-04 04:52:22.715656521 +0000 UTC m=+6554.007666233" Oct 04 04:52:31 crc kubenswrapper[4770]: I1004 04:52:31.762139 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-2xr9v"] Oct 04 04:52:31 crc kubenswrapper[4770]: I1004 04:52:31.763988 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2xr9v" Oct 04 04:52:31 crc kubenswrapper[4770]: I1004 04:52:31.785232 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2xr9v"] Oct 04 04:52:31 crc kubenswrapper[4770]: I1004 04:52:31.799700 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:52:31 crc kubenswrapper[4770]: I1004 04:52:31.799786 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:52:31 crc kubenswrapper[4770]: I1004 04:52:31.948965 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkznx\" (UniqueName: \"kubernetes.io/projected/9380ab45-8702-4380-aaae-31908686f28f-kube-api-access-fkznx\") pod \"manila-db-create-2xr9v\" (UID: \"9380ab45-8702-4380-aaae-31908686f28f\") " pod="openstack/manila-db-create-2xr9v" Oct 04 04:52:32 crc kubenswrapper[4770]: I1004 04:52:32.051237 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkznx\" (UniqueName: \"kubernetes.io/projected/9380ab45-8702-4380-aaae-31908686f28f-kube-api-access-fkznx\") pod \"manila-db-create-2xr9v\" (UID: \"9380ab45-8702-4380-aaae-31908686f28f\") " pod="openstack/manila-db-create-2xr9v" Oct 04 04:52:32 crc kubenswrapper[4770]: I1004 04:52:32.068735 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkznx\" (UniqueName: \"kubernetes.io/projected/9380ab45-8702-4380-aaae-31908686f28f-kube-api-access-fkznx\") pod \"manila-db-create-2xr9v\" (UID: \"9380ab45-8702-4380-aaae-31908686f28f\") " pod="openstack/manila-db-create-2xr9v" Oct 04 04:52:32 crc kubenswrapper[4770]: I1004 04:52:32.082266 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2xr9v" Oct 04 04:52:32 crc kubenswrapper[4770]: I1004 04:52:32.604862 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2xr9v"] Oct 04 04:52:32 crc kubenswrapper[4770]: I1004 04:52:32.802905 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2xr9v" event={"ID":"9380ab45-8702-4380-aaae-31908686f28f","Type":"ContainerStarted","Data":"98ef845cad1da717f46b31a0a46f2259962a9602704a666b4d3c9ea70de54b86"} Oct 04 04:52:33 crc kubenswrapper[4770]: I1004 04:52:33.813679 4770 generic.go:334] "Generic (PLEG): container finished" podID="9380ab45-8702-4380-aaae-31908686f28f" containerID="62e15a9865e1f8a92c17193409c2b4307c78d0c95f87813c067b2711cffcc250" exitCode=0 Oct 04 04:52:33 crc kubenswrapper[4770]: I1004 04:52:33.813774 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2xr9v" event={"ID":"9380ab45-8702-4380-aaae-31908686f28f","Type":"ContainerDied","Data":"62e15a9865e1f8a92c17193409c2b4307c78d0c95f87813c067b2711cffcc250"} Oct 04 04:52:35 crc kubenswrapper[4770]: I1004 04:52:35.407337 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2xr9v" Oct 04 04:52:35 crc kubenswrapper[4770]: I1004 04:52:35.527867 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkznx\" (UniqueName: \"kubernetes.io/projected/9380ab45-8702-4380-aaae-31908686f28f-kube-api-access-fkznx\") pod \"9380ab45-8702-4380-aaae-31908686f28f\" (UID: \"9380ab45-8702-4380-aaae-31908686f28f\") " Oct 04 04:52:35 crc kubenswrapper[4770]: I1004 04:52:35.538977 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9380ab45-8702-4380-aaae-31908686f28f-kube-api-access-fkznx" (OuterVolumeSpecName: "kube-api-access-fkznx") pod "9380ab45-8702-4380-aaae-31908686f28f" (UID: "9380ab45-8702-4380-aaae-31908686f28f"). InnerVolumeSpecName "kube-api-access-fkznx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:35 crc kubenswrapper[4770]: I1004 04:52:35.631710 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkznx\" (UniqueName: \"kubernetes.io/projected/9380ab45-8702-4380-aaae-31908686f28f-kube-api-access-fkznx\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:35 crc kubenswrapper[4770]: I1004 04:52:35.834329 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2xr9v" event={"ID":"9380ab45-8702-4380-aaae-31908686f28f","Type":"ContainerDied","Data":"98ef845cad1da717f46b31a0a46f2259962a9602704a666b4d3c9ea70de54b86"} Oct 04 04:52:35 crc kubenswrapper[4770]: I1004 04:52:35.834367 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98ef845cad1da717f46b31a0a46f2259962a9602704a666b4d3c9ea70de54b86" Oct 04 04:52:35 crc kubenswrapper[4770]: I1004 04:52:35.834449 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2xr9v" Oct 04 04:52:37 crc kubenswrapper[4770]: I1004 04:52:37.955689 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.033257 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-9178-account-create-4xk4p"] Oct 04 04:52:42 crc kubenswrapper[4770]: E1004 04:52:42.034311 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9380ab45-8702-4380-aaae-31908686f28f" containerName="mariadb-database-create" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.034326 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9380ab45-8702-4380-aaae-31908686f28f" containerName="mariadb-database-create" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.034527 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9380ab45-8702-4380-aaae-31908686f28f" containerName="mariadb-database-create" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.035318 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9178-account-create-4xk4p" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.037871 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.048511 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9178-account-create-4xk4p"] Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.163052 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lptrj\" (UniqueName: \"kubernetes.io/projected/7593e8ae-831e-4e1a-8ae0-81eea8df72b6-kube-api-access-lptrj\") pod \"manila-9178-account-create-4xk4p\" (UID: \"7593e8ae-831e-4e1a-8ae0-81eea8df72b6\") " pod="openstack/manila-9178-account-create-4xk4p" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.265159 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lptrj\" (UniqueName: \"kubernetes.io/projected/7593e8ae-831e-4e1a-8ae0-81eea8df72b6-kube-api-access-lptrj\") pod \"manila-9178-account-create-4xk4p\" (UID: \"7593e8ae-831e-4e1a-8ae0-81eea8df72b6\") " pod="openstack/manila-9178-account-create-4xk4p" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.292720 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lptrj\" (UniqueName: \"kubernetes.io/projected/7593e8ae-831e-4e1a-8ae0-81eea8df72b6-kube-api-access-lptrj\") pod \"manila-9178-account-create-4xk4p\" (UID: \"7593e8ae-831e-4e1a-8ae0-81eea8df72b6\") " pod="openstack/manila-9178-account-create-4xk4p" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.357201 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9178-account-create-4xk4p" Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.822363 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9178-account-create-4xk4p"] Oct 04 04:52:42 crc kubenswrapper[4770]: W1004 04:52:42.825319 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7593e8ae_831e_4e1a_8ae0_81eea8df72b6.slice/crio-f78016df08f761d43aced8ba3d7c3edb82932f8a47c03fc9d6409efc85a0382c WatchSource:0}: Error finding container f78016df08f761d43aced8ba3d7c3edb82932f8a47c03fc9d6409efc85a0382c: Status 404 returned error can't find the container with id f78016df08f761d43aced8ba3d7c3edb82932f8a47c03fc9d6409efc85a0382c Oct 04 04:52:42 crc kubenswrapper[4770]: I1004 04:52:42.914955 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9178-account-create-4xk4p" event={"ID":"7593e8ae-831e-4e1a-8ae0-81eea8df72b6","Type":"ContainerStarted","Data":"f78016df08f761d43aced8ba3d7c3edb82932f8a47c03fc9d6409efc85a0382c"} Oct 04 04:52:43 crc kubenswrapper[4770]: I1004 04:52:43.927051 4770 generic.go:334] "Generic (PLEG): container finished" podID="7593e8ae-831e-4e1a-8ae0-81eea8df72b6" containerID="f26c3eb368d2df57e8aae433521d2572669abe880afb4ab7eedaaed028d37ce3" exitCode=0 Oct 04 04:52:43 crc kubenswrapper[4770]: I1004 04:52:43.927316 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9178-account-create-4xk4p" event={"ID":"7593e8ae-831e-4e1a-8ae0-81eea8df72b6","Type":"ContainerDied","Data":"f26c3eb368d2df57e8aae433521d2572669abe880afb4ab7eedaaed028d37ce3"} Oct 04 04:52:45 crc kubenswrapper[4770]: I1004 04:52:45.337463 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9178-account-create-4xk4p" Oct 04 04:52:45 crc kubenswrapper[4770]: I1004 04:52:45.435176 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lptrj\" (UniqueName: \"kubernetes.io/projected/7593e8ae-831e-4e1a-8ae0-81eea8df72b6-kube-api-access-lptrj\") pod \"7593e8ae-831e-4e1a-8ae0-81eea8df72b6\" (UID: \"7593e8ae-831e-4e1a-8ae0-81eea8df72b6\") " Oct 04 04:52:45 crc kubenswrapper[4770]: I1004 04:52:45.441422 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7593e8ae-831e-4e1a-8ae0-81eea8df72b6-kube-api-access-lptrj" (OuterVolumeSpecName: "kube-api-access-lptrj") pod "7593e8ae-831e-4e1a-8ae0-81eea8df72b6" (UID: "7593e8ae-831e-4e1a-8ae0-81eea8df72b6"). InnerVolumeSpecName "kube-api-access-lptrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:45 crc kubenswrapper[4770]: I1004 04:52:45.538578 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lptrj\" (UniqueName: \"kubernetes.io/projected/7593e8ae-831e-4e1a-8ae0-81eea8df72b6-kube-api-access-lptrj\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:45 crc kubenswrapper[4770]: I1004 04:52:45.957135 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9178-account-create-4xk4p" event={"ID":"7593e8ae-831e-4e1a-8ae0-81eea8df72b6","Type":"ContainerDied","Data":"f78016df08f761d43aced8ba3d7c3edb82932f8a47c03fc9d6409efc85a0382c"} Oct 04 04:52:45 crc kubenswrapper[4770]: I1004 04:52:45.957214 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f78016df08f761d43aced8ba3d7c3edb82932f8a47c03fc9d6409efc85a0382c" Oct 04 04:52:45 crc kubenswrapper[4770]: I1004 04:52:45.957292 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9178-account-create-4xk4p" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.400882 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-ntrkf"] Oct 04 04:52:47 crc kubenswrapper[4770]: E1004 04:52:47.401933 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7593e8ae-831e-4e1a-8ae0-81eea8df72b6" containerName="mariadb-account-create" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.401947 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="7593e8ae-831e-4e1a-8ae0-81eea8df72b6" containerName="mariadb-account-create" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.402186 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="7593e8ae-831e-4e1a-8ae0-81eea8df72b6" containerName="mariadb-account-create" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.402962 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.405261 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.407289 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-7x5lf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.417448 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-ntrkf"] Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.482884 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-job-config-data\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.482949 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbk2c\" (UniqueName: \"kubernetes.io/projected/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-kube-api-access-wbk2c\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.483259 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-config-data\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.483376 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-combined-ca-bundle\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.585716 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-config-data\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.585786 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-combined-ca-bundle\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.585961 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-job-config-data\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.586657 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbk2c\" (UniqueName: \"kubernetes.io/projected/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-kube-api-access-wbk2c\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.591120 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-job-config-data\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.591360 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-combined-ca-bundle\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.592059 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-config-data\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.601819 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbk2c\" (UniqueName: \"kubernetes.io/projected/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-kube-api-access-wbk2c\") pod \"manila-db-sync-ntrkf\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:47 crc kubenswrapper[4770]: I1004 04:52:47.731880 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ntrkf" Oct 04 04:52:48 crc kubenswrapper[4770]: I1004 04:52:48.373709 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-ntrkf"] Oct 04 04:52:48 crc kubenswrapper[4770]: I1004 04:52:48.991592 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ntrkf" event={"ID":"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b","Type":"ContainerStarted","Data":"696fc006254260de7b0f49495fb5845678130e5c8cfeacc1dfb8e7d1a6221837"} Oct 04 04:53:01 crc kubenswrapper[4770]: I1004 04:53:01.795907 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:53:01 crc kubenswrapper[4770]: I1004 04:53:01.796476 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:53:01 crc kubenswrapper[4770]: I1004 04:53:01.796530 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:53:01 crc kubenswrapper[4770]: I1004 04:53:01.797398 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"284e8fec1a55dcb36285518859a2a4bc58d3161ca3219d7437f89da44e6444b6"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:53:01 crc kubenswrapper[4770]: I1004 04:53:01.797459 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://284e8fec1a55dcb36285518859a2a4bc58d3161ca3219d7437f89da44e6444b6" gracePeriod=600 Oct 04 04:53:02 crc kubenswrapper[4770]: I1004 04:53:02.139884 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ntrkf" event={"ID":"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b","Type":"ContainerStarted","Data":"6c97583d8738f2420c30f7f149d23254031d7f8ff8d974eb300b15f247d595ad"} Oct 04 04:53:02 crc kubenswrapper[4770]: I1004 04:53:02.143901 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="284e8fec1a55dcb36285518859a2a4bc58d3161ca3219d7437f89da44e6444b6" exitCode=0 Oct 04 04:53:02 crc kubenswrapper[4770]: I1004 04:53:02.143952 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"284e8fec1a55dcb36285518859a2a4bc58d3161ca3219d7437f89da44e6444b6"} Oct 04 04:53:02 crc kubenswrapper[4770]: I1004 04:53:02.143983 4770 scope.go:117] "RemoveContainer" containerID="272205eb3e262c157978565a7a021bacf179de6d0fa6702f7ca37bea5a250545" Oct 04 04:53:02 crc kubenswrapper[4770]: I1004 04:53:02.163851 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-ntrkf" podStartSLOduration=2.6225239670000002 podStartE2EDuration="15.16383266s" podCreationTimestamp="2025-10-04 04:52:47 +0000 UTC" firstStartedPulling="2025-10-04 04:52:48.383736986 +0000 UTC m=+6579.675746698" lastFinishedPulling="2025-10-04 04:53:00.925045679 +0000 UTC m=+6592.217055391" observedRunningTime="2025-10-04 04:53:02.15884827 +0000 UTC m=+6593.450857982" watchObservedRunningTime="2025-10-04 04:53:02.16383266 +0000 UTC m=+6593.455842372" Oct 04 04:53:03 crc kubenswrapper[4770]: I1004 04:53:03.154908 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac"} Oct 04 04:53:09 crc kubenswrapper[4770]: I1004 04:53:09.252159 4770 generic.go:334] "Generic (PLEG): container finished" podID="36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b" containerID="6c97583d8738f2420c30f7f149d23254031d7f8ff8d974eb300b15f247d595ad" exitCode=0 Oct 04 04:53:09 crc kubenswrapper[4770]: I1004 04:53:09.252266 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ntrkf" event={"ID":"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b","Type":"ContainerDied","Data":"6c97583d8738f2420c30f7f149d23254031d7f8ff8d974eb300b15f247d595ad"} Oct 04 04:53:10 crc kubenswrapper[4770]: I1004 04:53:10.783027 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ntrkf" Oct 04 04:53:10 crc kubenswrapper[4770]: I1004 04:53:10.914954 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbk2c\" (UniqueName: \"kubernetes.io/projected/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-kube-api-access-wbk2c\") pod \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " Oct 04 04:53:10 crc kubenswrapper[4770]: I1004 04:53:10.915029 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-combined-ca-bundle\") pod \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " Oct 04 04:53:10 crc kubenswrapper[4770]: I1004 04:53:10.915259 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-job-config-data\") pod \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " Oct 04 04:53:10 crc kubenswrapper[4770]: I1004 04:53:10.915351 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-config-data\") pod \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\" (UID: \"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b\") " Oct 04 04:53:10 crc kubenswrapper[4770]: I1004 04:53:10.922130 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-kube-api-access-wbk2c" (OuterVolumeSpecName: "kube-api-access-wbk2c") pod "36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b" (UID: "36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b"). InnerVolumeSpecName "kube-api-access-wbk2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.018407 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbk2c\" (UniqueName: \"kubernetes.io/projected/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-kube-api-access-wbk2c\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.018460 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b" (UID: "36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.020522 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-config-data" (OuterVolumeSpecName: "config-data") pod "36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b" (UID: "36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.020647 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b" (UID: "36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.119975 4770 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.120028 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.120038 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.278475 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ntrkf" event={"ID":"36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b","Type":"ContainerDied","Data":"696fc006254260de7b0f49495fb5845678130e5c8cfeacc1dfb8e7d1a6221837"} Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.278527 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="696fc006254260de7b0f49495fb5845678130e5c8cfeacc1dfb8e7d1a6221837" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.278553 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ntrkf" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.707950 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 04 04:53:11 crc kubenswrapper[4770]: E1004 04:53:11.708285 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b" containerName="manila-db-sync" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.708299 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b" containerName="manila-db-sync" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.708530 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b" containerName="manila-db-sync" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.709622 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.714323 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.717121 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.717272 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.717372 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-7x5lf" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.721070 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.760611 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.782548 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.782694 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.790948 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.844154 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6df44d9-b4kk8"] Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.853952 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.855826 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.855858 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-config-data\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.855890 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-config-data\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.855956 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e74b101f-cfa1-4c29-b78a-3c48964a271b-ceph\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.855980 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.855999 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e74b101f-cfa1-4c29-b78a-3c48964a271b-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.856057 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-scripts\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.856083 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.856100 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88256\" (UniqueName: \"kubernetes.io/projected/e74b101f-cfa1-4c29-b78a-3c48964a271b-kube-api-access-88256\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.856162 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp5mz\" (UniqueName: \"kubernetes.io/projected/7e858e72-afbf-45b7-96fb-c22def95cbed-kube-api-access-xp5mz\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.856180 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-scripts\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.856197 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/e74b101f-cfa1-4c29-b78a-3c48964a271b-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.856221 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.856240 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e858e72-afbf-45b7-96fb-c22def95cbed-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.865142 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6df44d9-b4kk8"] Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.958376 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-scripts\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.959827 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.959939 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88256\" (UniqueName: \"kubernetes.io/projected/e74b101f-cfa1-4c29-b78a-3c48964a271b-kube-api-access-88256\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.960157 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp5mz\" (UniqueName: \"kubernetes.io/projected/7e858e72-afbf-45b7-96fb-c22def95cbed-kube-api-access-xp5mz\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.960278 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-scripts\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.960373 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/e74b101f-cfa1-4c29-b78a-3c48964a271b-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.960470 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.960589 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-sb\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.960725 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e858e72-afbf-45b7-96fb-c22def95cbed-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.960843 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e858e72-afbf-45b7-96fb-c22def95cbed-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.960618 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/e74b101f-cfa1-4c29-b78a-3c48964a271b-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.961038 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvcr8\" (UniqueName: \"kubernetes.io/projected/529a8000-f96c-4d6e-9b47-28304e34975b-kube-api-access-rvcr8\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.961490 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-dns-svc\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.961585 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.961674 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-config-data\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.961809 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-config\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.961943 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-config-data\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.962213 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-nb\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.962964 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e74b101f-cfa1-4c29-b78a-3c48964a271b-ceph\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.964165 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.964271 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e74b101f-cfa1-4c29-b78a-3c48964a271b-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.964177 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-scripts\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.964545 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e74b101f-cfa1-4c29-b78a-3c48964a271b-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.964904 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.968204 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e74b101f-cfa1-4c29-b78a-3c48964a271b-ceph\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.968265 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.970861 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-config-data\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.972893 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-scripts\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.973888 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.976481 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74b101f-cfa1-4c29-b78a-3c48964a271b-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.982095 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88256\" (UniqueName: \"kubernetes.io/projected/e74b101f-cfa1-4c29-b78a-3c48964a271b-kube-api-access-88256\") pod \"manila-share-share1-0\" (UID: \"e74b101f-cfa1-4c29-b78a-3c48964a271b\") " pod="openstack/manila-share-share1-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.986345 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp5mz\" (UniqueName: \"kubernetes.io/projected/7e858e72-afbf-45b7-96fb-c22def95cbed-kube-api-access-xp5mz\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:11 crc kubenswrapper[4770]: I1004 04:53:11.987967 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e858e72-afbf-45b7-96fb-c22def95cbed-config-data\") pod \"manila-scheduler-0\" (UID: \"7e858e72-afbf-45b7-96fb-c22def95cbed\") " pod="openstack/manila-scheduler-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.005241 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.010720 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.012673 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.016746 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.065565 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066163 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-config-data-custom\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066203 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-logs\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066263 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-config-data\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066290 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-sb\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066403 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvcr8\" (UniqueName: \"kubernetes.io/projected/529a8000-f96c-4d6e-9b47-28304e34975b-kube-api-access-rvcr8\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066492 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-dns-svc\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066540 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066569 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7nnw\" (UniqueName: \"kubernetes.io/projected/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-kube-api-access-g7nnw\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066625 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-config\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066693 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-etc-machine-id\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066742 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-nb\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.066773 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-scripts\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.067249 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-sb\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.069247 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-dns-svc\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.069535 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-nb\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.069556 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-config\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.095441 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvcr8\" (UniqueName: \"kubernetes.io/projected/529a8000-f96c-4d6e-9b47-28304e34975b-kube-api-access-rvcr8\") pod \"dnsmasq-dns-7c6df44d9-b4kk8\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.133689 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.172550 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-config-data\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.172698 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.172731 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7nnw\" (UniqueName: \"kubernetes.io/projected/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-kube-api-access-g7nnw\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.172819 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-etc-machine-id\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.172860 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-scripts\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.172948 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-config-data-custom\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.172980 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-logs\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.173508 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-etc-machine-id\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.177136 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-logs\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.178554 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-config-data-custom\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.180624 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.182894 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-config-data\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.184795 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-scripts\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.201518 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.238650 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7nnw\" (UniqueName: \"kubernetes.io/projected/5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc-kube-api-access-g7nnw\") pod \"manila-api-0\" (UID: \"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc\") " pod="openstack/manila-api-0" Oct 04 04:53:12 crc kubenswrapper[4770]: I1004 04:53:12.260661 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 04 04:53:13 crc kubenswrapper[4770]: I1004 04:53:13.083731 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:53:13 crc kubenswrapper[4770]: I1004 04:53:13.104111 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 04 04:53:13 crc kubenswrapper[4770]: I1004 04:53:13.119436 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 04 04:53:13 crc kubenswrapper[4770]: I1004 04:53:13.201157 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6df44d9-b4kk8"] Oct 04 04:53:13 crc kubenswrapper[4770]: W1004 04:53:13.209326 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod529a8000_f96c_4d6e_9b47_28304e34975b.slice/crio-d8ebeba26b7954f80a2466e49b94cc0f19a3cee860fa461e2d31dfb8bc51b673 WatchSource:0}: Error finding container d8ebeba26b7954f80a2466e49b94cc0f19a3cee860fa461e2d31dfb8bc51b673: Status 404 returned error can't find the container with id d8ebeba26b7954f80a2466e49b94cc0f19a3cee860fa461e2d31dfb8bc51b673 Oct 04 04:53:13 crc kubenswrapper[4770]: I1004 04:53:13.367511 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" event={"ID":"529a8000-f96c-4d6e-9b47-28304e34975b","Type":"ContainerStarted","Data":"d8ebeba26b7954f80a2466e49b94cc0f19a3cee860fa461e2d31dfb8bc51b673"} Oct 04 04:53:13 crc kubenswrapper[4770]: I1004 04:53:13.370842 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"7e858e72-afbf-45b7-96fb-c22def95cbed","Type":"ContainerStarted","Data":"1228f24eae1c139157afe4e8a477e483badb1ea431c84a9490f79c0481c6091e"} Oct 04 04:53:13 crc kubenswrapper[4770]: I1004 04:53:13.381396 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"e74b101f-cfa1-4c29-b78a-3c48964a271b","Type":"ContainerStarted","Data":"9dc794438fe1cd0a3262506d6f7534fb8ee9eb4c28d7609c96d1bae34431f887"} Oct 04 04:53:13 crc kubenswrapper[4770]: I1004 04:53:13.384257 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 04 04:53:13 crc kubenswrapper[4770]: W1004 04:53:13.395637 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a9b426b_51b2_4e4f_bcb8_c0b84d6313dc.slice/crio-6c681e55cbf310fa9598aab75f9ace6d8279667b69e76ef95fd41ec24b1d9974 WatchSource:0}: Error finding container 6c681e55cbf310fa9598aab75f9ace6d8279667b69e76ef95fd41ec24b1d9974: Status 404 returned error can't find the container with id 6c681e55cbf310fa9598aab75f9ace6d8279667b69e76ef95fd41ec24b1d9974 Oct 04 04:53:14 crc kubenswrapper[4770]: I1004 04:53:14.411844 4770 generic.go:334] "Generic (PLEG): container finished" podID="529a8000-f96c-4d6e-9b47-28304e34975b" containerID="2a7238528e1861d770806a736fded11bf567371098e67fa2df0325247c78ac8a" exitCode=0 Oct 04 04:53:14 crc kubenswrapper[4770]: I1004 04:53:14.412371 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" event={"ID":"529a8000-f96c-4d6e-9b47-28304e34975b","Type":"ContainerDied","Data":"2a7238528e1861d770806a736fded11bf567371098e67fa2df0325247c78ac8a"} Oct 04 04:53:14 crc kubenswrapper[4770]: I1004 04:53:14.418513 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc","Type":"ContainerStarted","Data":"dfc27054ca4020ddc350afe608b600958c7f00289f51fc26b33555c96b936af4"} Oct 04 04:53:14 crc kubenswrapper[4770]: I1004 04:53:14.418581 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc","Type":"ContainerStarted","Data":"6c681e55cbf310fa9598aab75f9ace6d8279667b69e76ef95fd41ec24b1d9974"} Oct 04 04:53:15 crc kubenswrapper[4770]: I1004 04:53:15.440602 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc","Type":"ContainerStarted","Data":"f5cd759a80ef47e5e3b6b85ecf5b0a75295398706ae7f36682fda459be08884c"} Oct 04 04:53:15 crc kubenswrapper[4770]: I1004 04:53:15.441491 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 04 04:53:15 crc kubenswrapper[4770]: I1004 04:53:15.445717 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" event={"ID":"529a8000-f96c-4d6e-9b47-28304e34975b","Type":"ContainerStarted","Data":"7f956780180e7831b8b33a6ada5c9d2c9d20dddf030bf4525afdefaa97e2bb7e"} Oct 04 04:53:15 crc kubenswrapper[4770]: I1004 04:53:15.446136 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:15 crc kubenswrapper[4770]: I1004 04:53:15.447640 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"7e858e72-afbf-45b7-96fb-c22def95cbed","Type":"ContainerStarted","Data":"4276ddb9071c6d37c37cc47b14ee9b24d7c853fcedabe19554993f193a8e1c57"} Oct 04 04:53:15 crc kubenswrapper[4770]: I1004 04:53:15.474951 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.474928955 podStartE2EDuration="4.474928955s" podCreationTimestamp="2025-10-04 04:53:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:15.460394655 +0000 UTC m=+6606.752404367" watchObservedRunningTime="2025-10-04 04:53:15.474928955 +0000 UTC m=+6606.766938667" Oct 04 04:53:15 crc kubenswrapper[4770]: I1004 04:53:15.491299 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" podStartSLOduration=4.491274583 podStartE2EDuration="4.491274583s" podCreationTimestamp="2025-10-04 04:53:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:15.484910936 +0000 UTC m=+6606.776920658" watchObservedRunningTime="2025-10-04 04:53:15.491274583 +0000 UTC m=+6606.783284295" Oct 04 04:53:16 crc kubenswrapper[4770]: I1004 04:53:16.467311 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"7e858e72-afbf-45b7-96fb-c22def95cbed","Type":"ContainerStarted","Data":"d06f0c39cac123d48222e47ee267eac120f91c46466623682416e1aafb94d3e4"} Oct 04 04:53:19 crc kubenswrapper[4770]: I1004 04:53:19.713934 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=7.729964081 podStartE2EDuration="8.713914637s" podCreationTimestamp="2025-10-04 04:53:11 +0000 UTC" firstStartedPulling="2025-10-04 04:53:13.094634848 +0000 UTC m=+6604.386644560" lastFinishedPulling="2025-10-04 04:53:14.078585404 +0000 UTC m=+6605.370595116" observedRunningTime="2025-10-04 04:53:16.49442517 +0000 UTC m=+6607.786434882" watchObservedRunningTime="2025-10-04 04:53:19.713914637 +0000 UTC m=+6611.005924349" Oct 04 04:53:22 crc kubenswrapper[4770]: I1004 04:53:22.135360 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 04 04:53:22 crc kubenswrapper[4770]: I1004 04:53:22.203243 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:53:22 crc kubenswrapper[4770]: I1004 04:53:22.274141 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c66885869-nwzqr"] Oct 04 04:53:22 crc kubenswrapper[4770]: I1004 04:53:22.274398 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" podUID="6a94bbc1-1520-4811-b035-e812395b31a1" containerName="dnsmasq-dns" containerID="cri-o://2bed80eb41587a7333fdced63965c7b12a05d662454d30d64046d65f9a23f6eb" gracePeriod=10 Oct 04 04:53:24 crc kubenswrapper[4770]: I1004 04:53:24.554253 4770 generic.go:334] "Generic (PLEG): container finished" podID="6a94bbc1-1520-4811-b035-e812395b31a1" containerID="2bed80eb41587a7333fdced63965c7b12a05d662454d30d64046d65f9a23f6eb" exitCode=0 Oct 04 04:53:24 crc kubenswrapper[4770]: I1004 04:53:24.554386 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" event={"ID":"6a94bbc1-1520-4811-b035-e812395b31a1","Type":"ContainerDied","Data":"2bed80eb41587a7333fdced63965c7b12a05d662454d30d64046d65f9a23f6eb"} Oct 04 04:53:26 crc kubenswrapper[4770]: I1004 04:53:26.021941 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" podUID="6a94bbc1-1520-4811-b035-e812395b31a1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.79:5353: connect: connection refused" Oct 04 04:53:26 crc kubenswrapper[4770]: I1004 04:53:26.941401 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.060373 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch9vd\" (UniqueName: \"kubernetes.io/projected/6a94bbc1-1520-4811-b035-e812395b31a1-kube-api-access-ch9vd\") pod \"6a94bbc1-1520-4811-b035-e812395b31a1\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.061061 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-dns-svc\") pod \"6a94bbc1-1520-4811-b035-e812395b31a1\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.061226 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-sb\") pod \"6a94bbc1-1520-4811-b035-e812395b31a1\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.061311 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-nb\") pod \"6a94bbc1-1520-4811-b035-e812395b31a1\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.061426 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-config\") pod \"6a94bbc1-1520-4811-b035-e812395b31a1\" (UID: \"6a94bbc1-1520-4811-b035-e812395b31a1\") " Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.097785 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-tr7wv"] Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.103605 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a94bbc1-1520-4811-b035-e812395b31a1-kube-api-access-ch9vd" (OuterVolumeSpecName: "kube-api-access-ch9vd") pod "6a94bbc1-1520-4811-b035-e812395b31a1" (UID: "6a94bbc1-1520-4811-b035-e812395b31a1"). InnerVolumeSpecName "kube-api-access-ch9vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.131786 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-tr7wv"] Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.166884 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch9vd\" (UniqueName: \"kubernetes.io/projected/6a94bbc1-1520-4811-b035-e812395b31a1-kube-api-access-ch9vd\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.183870 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-config" (OuterVolumeSpecName: "config") pod "6a94bbc1-1520-4811-b035-e812395b31a1" (UID: "6a94bbc1-1520-4811-b035-e812395b31a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.201599 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6a94bbc1-1520-4811-b035-e812395b31a1" (UID: "6a94bbc1-1520-4811-b035-e812395b31a1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.247639 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6a94bbc1-1520-4811-b035-e812395b31a1" (UID: "6a94bbc1-1520-4811-b035-e812395b31a1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.259886 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a94bbc1-1520-4811-b035-e812395b31a1" (UID: "6a94bbc1-1520-4811-b035-e812395b31a1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.274342 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.274593 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.274690 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.274766 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a94bbc1-1520-4811-b035-e812395b31a1-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.560626 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.561285 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="ceilometer-central-agent" containerID="cri-o://a1079219452f4179d5885f881c017c81f1500a95066d83c2484944ee35e48936" gracePeriod=30 Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.561338 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="proxy-httpd" containerID="cri-o://b5b3c8a655fe4f33994b15d5d0f43f87d9564556eb7e01563c59a9cf9c10c247" gracePeriod=30 Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.561435 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="sg-core" containerID="cri-o://51894a8e3ad90a2fdd772ce71ebc448c05a9fc853d32f2fbf169b5f4bb01ec54" gracePeriod=30 Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.561482 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="ceilometer-notification-agent" containerID="cri-o://1f00cb8a22cc527ca99b4475e123754e10b32963dcc5c396be7763e9230a4a78" gracePeriod=30 Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.602709 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"e74b101f-cfa1-4c29-b78a-3c48964a271b","Type":"ContainerStarted","Data":"2e959df056e37510854c49f04e26bfd258419e01cd4f9b866d6f3212a1286681"} Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.607435 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" event={"ID":"6a94bbc1-1520-4811-b035-e812395b31a1","Type":"ContainerDied","Data":"c5f03a128ab433fd913e8d9aac8747b2ccf6c209e9ca61b53cfaeaee6995a0b0"} Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.607502 4770 scope.go:117] "RemoveContainer" containerID="2bed80eb41587a7333fdced63965c7b12a05d662454d30d64046d65f9a23f6eb" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.607675 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.697110 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74462d5e-59ea-4724-bdd5-d68f3640a0ca" path="/var/lib/kubelet/pods/74462d5e-59ea-4724-bdd5-d68f3640a0ca/volumes" Oct 04 04:53:27 crc kubenswrapper[4770]: I1004 04:53:27.759435 4770 scope.go:117] "RemoveContainer" containerID="2537f03ad69d69d119ae0d33630a3e794fe8a004ad722e801026ef7c22da7135" Oct 04 04:53:28 crc kubenswrapper[4770]: I1004 04:53:28.620003 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"e74b101f-cfa1-4c29-b78a-3c48964a271b","Type":"ContainerStarted","Data":"c325fe236baac4478755bc34b7657f3b21d955ba4949a8020bd396a09dad6aba"} Oct 04 04:53:28 crc kubenswrapper[4770]: I1004 04:53:28.626520 4770 generic.go:334] "Generic (PLEG): container finished" podID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerID="b5b3c8a655fe4f33994b15d5d0f43f87d9564556eb7e01563c59a9cf9c10c247" exitCode=0 Oct 04 04:53:28 crc kubenswrapper[4770]: I1004 04:53:28.626559 4770 generic.go:334] "Generic (PLEG): container finished" podID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerID="51894a8e3ad90a2fdd772ce71ebc448c05a9fc853d32f2fbf169b5f4bb01ec54" exitCode=2 Oct 04 04:53:28 crc kubenswrapper[4770]: I1004 04:53:28.626574 4770 generic.go:334] "Generic (PLEG): container finished" podID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerID="a1079219452f4179d5885f881c017c81f1500a95066d83c2484944ee35e48936" exitCode=0 Oct 04 04:53:28 crc kubenswrapper[4770]: I1004 04:53:28.626603 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerDied","Data":"b5b3c8a655fe4f33994b15d5d0f43f87d9564556eb7e01563c59a9cf9c10c247"} Oct 04 04:53:28 crc kubenswrapper[4770]: I1004 04:53:28.626632 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerDied","Data":"51894a8e3ad90a2fdd772ce71ebc448c05a9fc853d32f2fbf169b5f4bb01ec54"} Oct 04 04:53:28 crc kubenswrapper[4770]: I1004 04:53:28.626653 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerDied","Data":"a1079219452f4179d5885f881c017c81f1500a95066d83c2484944ee35e48936"} Oct 04 04:53:28 crc kubenswrapper[4770]: I1004 04:53:28.648900 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.808146335 podStartE2EDuration="17.648875643s" podCreationTimestamp="2025-10-04 04:53:11 +0000 UTC" firstStartedPulling="2025-10-04 04:53:13.083434325 +0000 UTC m=+6604.375444037" lastFinishedPulling="2025-10-04 04:53:26.924163633 +0000 UTC m=+6618.216173345" observedRunningTime="2025-10-04 04:53:28.640962817 +0000 UTC m=+6619.932972539" watchObservedRunningTime="2025-10-04 04:53:28.648875643 +0000 UTC m=+6619.940885355" Oct 04 04:53:29 crc kubenswrapper[4770]: I1004 04:53:29.162059 4770 scope.go:117] "RemoveContainer" containerID="f36650878e20c4c6bc39791f4ad25ad88d047565db4069f8ad0821914e9629cd" Oct 04 04:53:30 crc kubenswrapper[4770]: E1004 04:53:30.280962 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc57c72ce_d6a9_45fd_a0e6_ccd7b974380f.slice/crio-1f00cb8a22cc527ca99b4475e123754e10b32963dcc5c396be7763e9230a4a78.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.650523 4770 generic.go:334] "Generic (PLEG): container finished" podID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerID="1f00cb8a22cc527ca99b4475e123754e10b32963dcc5c396be7763e9230a4a78" exitCode=0 Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.650572 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerDied","Data":"1f00cb8a22cc527ca99b4475e123754e10b32963dcc5c396be7763e9230a4a78"} Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.744071 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.850350 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-run-httpd\") pod \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.850402 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-config-data\") pod \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.850425 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-combined-ca-bundle\") pod \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.850544 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-log-httpd\") pod \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.850597 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-sg-core-conf-yaml\") pod \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.850640 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-scripts\") pod \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.850673 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s2c2\" (UniqueName: \"kubernetes.io/projected/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-kube-api-access-2s2c2\") pod \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\" (UID: \"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f\") " Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.852719 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" (UID: "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.854021 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" (UID: "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.861358 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-kube-api-access-2s2c2" (OuterVolumeSpecName: "kube-api-access-2s2c2") pod "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" (UID: "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f"). InnerVolumeSpecName "kube-api-access-2s2c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.861368 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-scripts" (OuterVolumeSpecName: "scripts") pod "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" (UID: "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.886575 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" (UID: "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.951163 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" (UID: "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.952514 4770 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.952538 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.952552 4770 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.952563 4770 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.952574 4770 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.952585 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s2c2\" (UniqueName: \"kubernetes.io/projected/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-kube-api-access-2s2c2\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:30 crc kubenswrapper[4770]: I1004 04:53:30.968159 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-config-data" (OuterVolumeSpecName: "config-data") pod "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" (UID: "c57c72ce-d6a9-45fd-a0e6-ccd7b974380f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.054354 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.670672 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c57c72ce-d6a9-45fd-a0e6-ccd7b974380f","Type":"ContainerDied","Data":"a048cafc866b13a2262cc6bb0a9cc80a314a4fc895fd81b7a7812eac720684f1"} Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.670751 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.671823 4770 scope.go:117] "RemoveContainer" containerID="b5b3c8a655fe4f33994b15d5d0f43f87d9564556eb7e01563c59a9cf9c10c247" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.708529 4770 scope.go:117] "RemoveContainer" containerID="51894a8e3ad90a2fdd772ce71ebc448c05a9fc853d32f2fbf169b5f4bb01ec54" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.730199 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.747290 4770 scope.go:117] "RemoveContainer" containerID="1f00cb8a22cc527ca99b4475e123754e10b32963dcc5c396be7763e9230a4a78" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.751725 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.763801 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:31 crc kubenswrapper[4770]: E1004 04:53:31.764420 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="proxy-httpd" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764439 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="proxy-httpd" Oct 04 04:53:31 crc kubenswrapper[4770]: E1004 04:53:31.764475 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="ceilometer-central-agent" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764488 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="ceilometer-central-agent" Oct 04 04:53:31 crc kubenswrapper[4770]: E1004 04:53:31.764520 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a94bbc1-1520-4811-b035-e812395b31a1" containerName="dnsmasq-dns" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764530 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a94bbc1-1520-4811-b035-e812395b31a1" containerName="dnsmasq-dns" Oct 04 04:53:31 crc kubenswrapper[4770]: E1004 04:53:31.764540 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a94bbc1-1520-4811-b035-e812395b31a1" containerName="init" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764548 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a94bbc1-1520-4811-b035-e812395b31a1" containerName="init" Oct 04 04:53:31 crc kubenswrapper[4770]: E1004 04:53:31.764573 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="sg-core" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764581 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="sg-core" Oct 04 04:53:31 crc kubenswrapper[4770]: E1004 04:53:31.764607 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="ceilometer-notification-agent" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764616 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="ceilometer-notification-agent" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764862 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="ceilometer-notification-agent" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764885 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a94bbc1-1520-4811-b035-e812395b31a1" containerName="dnsmasq-dns" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764906 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="ceilometer-central-agent" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764935 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="proxy-httpd" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.764946 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" containerName="sg-core" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.768473 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.773552 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.775281 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.775282 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.775378 4770 scope.go:117] "RemoveContainer" containerID="a1079219452f4179d5885f881c017c81f1500a95066d83c2484944ee35e48936" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.871507 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7631708-8285-44c1-b416-ad9e0dae57b8-log-httpd\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.871703 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.871758 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-config-data\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.871994 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7631708-8285-44c1-b416-ad9e0dae57b8-run-httpd\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.872155 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-scripts\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.872221 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpdwc\" (UniqueName: \"kubernetes.io/projected/c7631708-8285-44c1-b416-ad9e0dae57b8-kube-api-access-zpdwc\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.872348 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.974705 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.974751 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-config-data\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.974822 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7631708-8285-44c1-b416-ad9e0dae57b8-run-httpd\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.974864 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-scripts\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.974897 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpdwc\" (UniqueName: \"kubernetes.io/projected/c7631708-8285-44c1-b416-ad9e0dae57b8-kube-api-access-zpdwc\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.974936 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.974976 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7631708-8285-44c1-b416-ad9e0dae57b8-log-httpd\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.975416 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7631708-8285-44c1-b416-ad9e0dae57b8-log-httpd\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.976452 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7631708-8285-44c1-b416-ad9e0dae57b8-run-httpd\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.979467 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-scripts\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.980400 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-config-data\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.980522 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:31 crc kubenswrapper[4770]: I1004 04:53:31.981916 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7631708-8285-44c1-b416-ad9e0dae57b8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:32 crc kubenswrapper[4770]: I1004 04:53:32.007336 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpdwc\" (UniqueName: \"kubernetes.io/projected/c7631708-8285-44c1-b416-ad9e0dae57b8-kube-api-access-zpdwc\") pod \"ceilometer-0\" (UID: \"c7631708-8285-44c1-b416-ad9e0dae57b8\") " pod="openstack/ceilometer-0" Oct 04 04:53:32 crc kubenswrapper[4770]: I1004 04:53:32.066293 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 04 04:53:32 crc kubenswrapper[4770]: I1004 04:53:32.108061 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:53:32 crc kubenswrapper[4770]: I1004 04:53:32.585148 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:32 crc kubenswrapper[4770]: I1004 04:53:32.686303 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7631708-8285-44c1-b416-ad9e0dae57b8","Type":"ContainerStarted","Data":"ca0847d9e0e419ab15a1a21fd1a00ace31cfb61eec68631bd4ba22a37a1a6894"} Oct 04 04:53:33 crc kubenswrapper[4770]: I1004 04:53:33.596454 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 04 04:53:33 crc kubenswrapper[4770]: I1004 04:53:33.689839 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c57c72ce-d6a9-45fd-a0e6-ccd7b974380f" path="/var/lib/kubelet/pods/c57c72ce-d6a9-45fd-a0e6-ccd7b974380f/volumes" Oct 04 04:53:33 crc kubenswrapper[4770]: I1004 04:53:33.793865 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 04 04:53:34 crc kubenswrapper[4770]: I1004 04:53:34.707902 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7631708-8285-44c1-b416-ad9e0dae57b8","Type":"ContainerStarted","Data":"fe9a0550da84c55c86a5d6a98ea93e22e772bdf90a1a4c599a912b83bda57a12"} Oct 04 04:53:36 crc kubenswrapper[4770]: I1004 04:53:36.728913 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7631708-8285-44c1-b416-ad9e0dae57b8","Type":"ContainerStarted","Data":"f09ab7e127c7504403ba56aea5422f58f6f069e2685e246fefecef645b54f32c"} Oct 04 04:53:38 crc kubenswrapper[4770]: I1004 04:53:38.758395 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7631708-8285-44c1-b416-ad9e0dae57b8","Type":"ContainerStarted","Data":"62556d825b08c8add9c060e481398a8b8ecd0d8b400c2646f21c08ed5cb85d1a"} Oct 04 04:53:40 crc kubenswrapper[4770]: I1004 04:53:40.038115 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-f8cf-account-create-cst6w"] Oct 04 04:53:40 crc kubenswrapper[4770]: I1004 04:53:40.047869 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-f8cf-account-create-cst6w"] Oct 04 04:53:41 crc kubenswrapper[4770]: I1004 04:53:41.693976 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d300c24-108b-4b17-a84c-035a0db8b6bc" path="/var/lib/kubelet/pods/7d300c24-108b-4b17-a84c-035a0db8b6bc/volumes" Oct 04 04:53:43 crc kubenswrapper[4770]: I1004 04:53:43.651212 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 04 04:53:45 crc kubenswrapper[4770]: I1004 04:53:45.845052 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7631708-8285-44c1-b416-ad9e0dae57b8","Type":"ContainerStarted","Data":"dd67de463e1dfa706612e894706cc986ed945d5784a0ae5a387fb2510d89a7d3"} Oct 04 04:53:45 crc kubenswrapper[4770]: I1004 04:53:45.845962 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:53:45 crc kubenswrapper[4770]: I1004 04:53:45.872908 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.07975415 podStartE2EDuration="14.872889643s" podCreationTimestamp="2025-10-04 04:53:31 +0000 UTC" firstStartedPulling="2025-10-04 04:53:32.595046307 +0000 UTC m=+6623.887056019" lastFinishedPulling="2025-10-04 04:53:44.38818179 +0000 UTC m=+6635.680191512" observedRunningTime="2025-10-04 04:53:45.866561887 +0000 UTC m=+6637.158571609" watchObservedRunningTime="2025-10-04 04:53:45.872889643 +0000 UTC m=+6637.164899355" Oct 04 04:53:46 crc kubenswrapper[4770]: I1004 04:53:46.025956 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-5kgpr"] Oct 04 04:53:46 crc kubenswrapper[4770]: I1004 04:53:46.034989 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-5kgpr"] Oct 04 04:53:47 crc kubenswrapper[4770]: I1004 04:53:47.695620 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1" path="/var/lib/kubelet/pods/bca2a7b2-ae4d-4fb5-88df-9b6dfd4165e1/volumes" Oct 04 04:53:57 crc kubenswrapper[4770]: I1004 04:53:57.052656 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-da4f-account-create-l27vj"] Oct 04 04:53:57 crc kubenswrapper[4770]: I1004 04:53:57.064955 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-da4f-account-create-l27vj"] Oct 04 04:53:57 crc kubenswrapper[4770]: I1004 04:53:57.683198 4770 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod6a94bbc1-1520-4811-b035-e812395b31a1"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod6a94bbc1-1520-4811-b035-e812395b31a1] : Timed out while waiting for systemd to remove kubepods-besteffort-pod6a94bbc1_1520_4811_b035_e812395b31a1.slice" Oct 04 04:53:57 crc kubenswrapper[4770]: E1004 04:53:57.683247 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod6a94bbc1-1520-4811-b035-e812395b31a1] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod6a94bbc1-1520-4811-b035-e812395b31a1] : Timed out while waiting for systemd to remove kubepods-besteffort-pod6a94bbc1_1520_4811_b035_e812395b31a1.slice" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" podUID="6a94bbc1-1520-4811-b035-e812395b31a1" Oct 04 04:53:57 crc kubenswrapper[4770]: I1004 04:53:57.687216 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd207fcd-44df-4b1b-bff9-630f556d7a5f" path="/var/lib/kubelet/pods/dd207fcd-44df-4b1b-bff9-630f556d7a5f/volumes" Oct 04 04:53:57 crc kubenswrapper[4770]: I1004 04:53:57.975396 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c66885869-nwzqr" Oct 04 04:53:58 crc kubenswrapper[4770]: I1004 04:53:58.021419 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c66885869-nwzqr"] Oct 04 04:53:58 crc kubenswrapper[4770]: I1004 04:53:58.032271 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c66885869-nwzqr"] Oct 04 04:53:59 crc kubenswrapper[4770]: I1004 04:53:59.686032 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a94bbc1-1520-4811-b035-e812395b31a1" path="/var/lib/kubelet/pods/6a94bbc1-1520-4811-b035-e812395b31a1/volumes" Oct 04 04:54:02 crc kubenswrapper[4770]: I1004 04:54:02.115325 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 04:54:29 crc kubenswrapper[4770]: I1004 04:54:29.354842 4770 scope.go:117] "RemoveContainer" containerID="0b92cd5705bee2cb7dc44d7c45166bdb8831b73d134ca7f6985bb754df3f1f86" Oct 04 04:54:29 crc kubenswrapper[4770]: I1004 04:54:29.395178 4770 scope.go:117] "RemoveContainer" containerID="5e322bda6df305b40256f3b062dc6710756a3c6fd9a6133f1a72bfd80be65ab3" Oct 04 04:54:29 crc kubenswrapper[4770]: I1004 04:54:29.667886 4770 scope.go:117] "RemoveContainer" containerID="d6f4a444c09ba05fa6e129d32ef43d4320ee28d7ba3b0b8a19592d9da8c7accb" Oct 04 04:54:29 crc kubenswrapper[4770]: I1004 04:54:29.698162 4770 scope.go:117] "RemoveContainer" containerID="cce914e63f656abc95c2d67af215f914022cec3d6c94b3b1ca8ee891fd5fb80f" Oct 04 04:54:29 crc kubenswrapper[4770]: I1004 04:54:29.758220 4770 scope.go:117] "RemoveContainer" containerID="ff2b58e303ddc23248e6155a5eae386811b9a87ada716158f0026065b9474e5a" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.556053 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64c78898cc-v6nfg"] Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.558455 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.561561 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.580837 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64c78898cc-v6nfg"] Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.718385 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-config\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.718493 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hx47\" (UniqueName: \"kubernetes.io/projected/0f2de188-5276-445c-a166-c4ffd0a68380-kube-api-access-7hx47\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.718534 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-dns-svc\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.718645 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-nb\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.718710 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-openstack-cell1\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.718767 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-sb\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.820947 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-nb\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.821297 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-openstack-cell1\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.821336 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-sb\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.821402 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-config\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.821497 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hx47\" (UniqueName: \"kubernetes.io/projected/0f2de188-5276-445c-a166-c4ffd0a68380-kube-api-access-7hx47\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.821541 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-dns-svc\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.823379 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-sb\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.823817 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-dns-svc\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.823823 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-openstack-cell1\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.823817 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-config\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.824236 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-nb\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.841788 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hx47\" (UniqueName: \"kubernetes.io/projected/0f2de188-5276-445c-a166-c4ffd0a68380-kube-api-access-7hx47\") pod \"dnsmasq-dns-64c78898cc-v6nfg\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:31 crc kubenswrapper[4770]: I1004 04:54:31.889104 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:32 crc kubenswrapper[4770]: I1004 04:54:32.515483 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64c78898cc-v6nfg"] Oct 04 04:54:33 crc kubenswrapper[4770]: I1004 04:54:33.372394 4770 generic.go:334] "Generic (PLEG): container finished" podID="0f2de188-5276-445c-a166-c4ffd0a68380" containerID="568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b" exitCode=0 Oct 04 04:54:33 crc kubenswrapper[4770]: I1004 04:54:33.372584 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" event={"ID":"0f2de188-5276-445c-a166-c4ffd0a68380","Type":"ContainerDied","Data":"568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b"} Oct 04 04:54:33 crc kubenswrapper[4770]: I1004 04:54:33.372964 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" event={"ID":"0f2de188-5276-445c-a166-c4ffd0a68380","Type":"ContainerStarted","Data":"715cd77a04f8de78dc488b154f1543dcd175697795e7f78071913b54a7551bf7"} Oct 04 04:54:34 crc kubenswrapper[4770]: I1004 04:54:34.386633 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" event={"ID":"0f2de188-5276-445c-a166-c4ffd0a68380","Type":"ContainerStarted","Data":"4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25"} Oct 04 04:54:34 crc kubenswrapper[4770]: I1004 04:54:34.387279 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:34 crc kubenswrapper[4770]: I1004 04:54:34.411613 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" podStartSLOduration=3.411599238 podStartE2EDuration="3.411599238s" podCreationTimestamp="2025-10-04 04:54:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:34.408109136 +0000 UTC m=+6685.700118848" watchObservedRunningTime="2025-10-04 04:54:34.411599238 +0000 UTC m=+6685.703608950" Oct 04 04:54:41 crc kubenswrapper[4770]: I1004 04:54:41.890722 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:41 crc kubenswrapper[4770]: I1004 04:54:41.966432 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6df44d9-b4kk8"] Oct 04 04:54:41 crc kubenswrapper[4770]: I1004 04:54:41.966702 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" podUID="529a8000-f96c-4d6e-9b47-28304e34975b" containerName="dnsmasq-dns" containerID="cri-o://7f956780180e7831b8b33a6ada5c9d2c9d20dddf030bf4525afdefaa97e2bb7e" gracePeriod=10 Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.202952 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" podUID="529a8000-f96c-4d6e-9b47-28304e34975b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.146:5353: connect: connection refused" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.239484 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f684d669f-jkbvx"] Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.255966 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.273875 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f684d669f-jkbvx"] Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.375500 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-dns-svc\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.375769 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-ovsdbserver-nb\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.376185 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp2xf\" (UniqueName: \"kubernetes.io/projected/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-kube-api-access-sp2xf\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.376457 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-ovsdbserver-sb\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.376583 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-openstack-cell1\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.376837 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-config\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.471891 4770 generic.go:334] "Generic (PLEG): container finished" podID="529a8000-f96c-4d6e-9b47-28304e34975b" containerID="7f956780180e7831b8b33a6ada5c9d2c9d20dddf030bf4525afdefaa97e2bb7e" exitCode=0 Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.471938 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" event={"ID":"529a8000-f96c-4d6e-9b47-28304e34975b","Type":"ContainerDied","Data":"7f956780180e7831b8b33a6ada5c9d2c9d20dddf030bf4525afdefaa97e2bb7e"} Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.478122 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-ovsdbserver-nb\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.478190 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp2xf\" (UniqueName: \"kubernetes.io/projected/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-kube-api-access-sp2xf\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.478275 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-ovsdbserver-sb\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.478321 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-openstack-cell1\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.478366 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-config\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.478385 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-dns-svc\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.479328 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-ovsdbserver-nb\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.479334 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-openstack-cell1\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.479443 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-dns-svc\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.479521 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-config\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.480721 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-ovsdbserver-sb\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.497128 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp2xf\" (UniqueName: \"kubernetes.io/projected/cc184b0a-0bf4-4a76-b428-02d3e66e7eb2-kube-api-access-sp2xf\") pod \"dnsmasq-dns-f684d669f-jkbvx\" (UID: \"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2\") " pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:42 crc kubenswrapper[4770]: I1004 04:54:42.599867 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.189484 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.244475 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f684d669f-jkbvx"] Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.328474 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-config\") pod \"529a8000-f96c-4d6e-9b47-28304e34975b\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.328561 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-sb\") pod \"529a8000-f96c-4d6e-9b47-28304e34975b\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.328654 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-nb\") pod \"529a8000-f96c-4d6e-9b47-28304e34975b\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.328764 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvcr8\" (UniqueName: \"kubernetes.io/projected/529a8000-f96c-4d6e-9b47-28304e34975b-kube-api-access-rvcr8\") pod \"529a8000-f96c-4d6e-9b47-28304e34975b\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.328926 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-dns-svc\") pod \"529a8000-f96c-4d6e-9b47-28304e34975b\" (UID: \"529a8000-f96c-4d6e-9b47-28304e34975b\") " Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.351479 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/529a8000-f96c-4d6e-9b47-28304e34975b-kube-api-access-rvcr8" (OuterVolumeSpecName: "kube-api-access-rvcr8") pod "529a8000-f96c-4d6e-9b47-28304e34975b" (UID: "529a8000-f96c-4d6e-9b47-28304e34975b"). InnerVolumeSpecName "kube-api-access-rvcr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.403368 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "529a8000-f96c-4d6e-9b47-28304e34975b" (UID: "529a8000-f96c-4d6e-9b47-28304e34975b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.418677 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "529a8000-f96c-4d6e-9b47-28304e34975b" (UID: "529a8000-f96c-4d6e-9b47-28304e34975b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.422344 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "529a8000-f96c-4d6e-9b47-28304e34975b" (UID: "529a8000-f96c-4d6e-9b47-28304e34975b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.423531 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-config" (OuterVolumeSpecName: "config") pod "529a8000-f96c-4d6e-9b47-28304e34975b" (UID: "529a8000-f96c-4d6e-9b47-28304e34975b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.433055 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvcr8\" (UniqueName: \"kubernetes.io/projected/529a8000-f96c-4d6e-9b47-28304e34975b-kube-api-access-rvcr8\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.433083 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.433114 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.433125 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.433134 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/529a8000-f96c-4d6e-9b47-28304e34975b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.483361 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" event={"ID":"529a8000-f96c-4d6e-9b47-28304e34975b","Type":"ContainerDied","Data":"d8ebeba26b7954f80a2466e49b94cc0f19a3cee860fa461e2d31dfb8bc51b673"} Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.483686 4770 scope.go:117] "RemoveContainer" containerID="7f956780180e7831b8b33a6ada5c9d2c9d20dddf030bf4525afdefaa97e2bb7e" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.483390 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6df44d9-b4kk8" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.486036 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f684d669f-jkbvx" event={"ID":"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2","Type":"ContainerStarted","Data":"0847e6c599451a3554a1d7b8b723bdfe304eea997086654a06b692d84eefe7f7"} Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.520434 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6df44d9-b4kk8"] Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.521425 4770 scope.go:117] "RemoveContainer" containerID="2a7238528e1861d770806a736fded11bf567371098e67fa2df0325247c78ac8a" Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.529301 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6df44d9-b4kk8"] Oct 04 04:54:43 crc kubenswrapper[4770]: I1004 04:54:43.692694 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="529a8000-f96c-4d6e-9b47-28304e34975b" path="/var/lib/kubelet/pods/529a8000-f96c-4d6e-9b47-28304e34975b/volumes" Oct 04 04:54:44 crc kubenswrapper[4770]: I1004 04:54:44.496996 4770 generic.go:334] "Generic (PLEG): container finished" podID="cc184b0a-0bf4-4a76-b428-02d3e66e7eb2" containerID="f5fcde5b65b3448f7a364227cf1d1184019e11b23ca7affd3ecb6323f5a7953a" exitCode=0 Oct 04 04:54:44 crc kubenswrapper[4770]: I1004 04:54:44.497235 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f684d669f-jkbvx" event={"ID":"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2","Type":"ContainerDied","Data":"f5fcde5b65b3448f7a364227cf1d1184019e11b23ca7affd3ecb6323f5a7953a"} Oct 04 04:54:45 crc kubenswrapper[4770]: I1004 04:54:45.512607 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f684d669f-jkbvx" event={"ID":"cc184b0a-0bf4-4a76-b428-02d3e66e7eb2","Type":"ContainerStarted","Data":"c205cb96047f5eb08ca404156e4b8a03695a559f1ff0a20f67469518a3cc616b"} Oct 04 04:54:45 crc kubenswrapper[4770]: I1004 04:54:45.512897 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:45 crc kubenswrapper[4770]: I1004 04:54:45.539915 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f684d669f-jkbvx" podStartSLOduration=3.539891132 podStartE2EDuration="3.539891132s" podCreationTimestamp="2025-10-04 04:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:45.533239338 +0000 UTC m=+6696.825249120" watchObservedRunningTime="2025-10-04 04:54:45.539891132 +0000 UTC m=+6696.831900854" Oct 04 04:54:52 crc kubenswrapper[4770]: I1004 04:54:52.601927 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f684d669f-jkbvx" Oct 04 04:54:52 crc kubenswrapper[4770]: I1004 04:54:52.694641 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64c78898cc-v6nfg"] Oct 04 04:54:52 crc kubenswrapper[4770]: I1004 04:54:52.694881 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" podUID="0f2de188-5276-445c-a166-c4ffd0a68380" containerName="dnsmasq-dns" containerID="cri-o://4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25" gracePeriod=10 Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.481799 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.607401 4770 generic.go:334] "Generic (PLEG): container finished" podID="0f2de188-5276-445c-a166-c4ffd0a68380" containerID="4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25" exitCode=0 Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.607461 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" event={"ID":"0f2de188-5276-445c-a166-c4ffd0a68380","Type":"ContainerDied","Data":"4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25"} Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.607481 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.607504 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c78898cc-v6nfg" event={"ID":"0f2de188-5276-445c-a166-c4ffd0a68380","Type":"ContainerDied","Data":"715cd77a04f8de78dc488b154f1543dcd175697795e7f78071913b54a7551bf7"} Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.607528 4770 scope.go:117] "RemoveContainer" containerID="4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.611849 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hx47\" (UniqueName: \"kubernetes.io/projected/0f2de188-5276-445c-a166-c4ffd0a68380-kube-api-access-7hx47\") pod \"0f2de188-5276-445c-a166-c4ffd0a68380\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.611936 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-openstack-cell1\") pod \"0f2de188-5276-445c-a166-c4ffd0a68380\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.611960 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-sb\") pod \"0f2de188-5276-445c-a166-c4ffd0a68380\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.612048 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-nb\") pod \"0f2de188-5276-445c-a166-c4ffd0a68380\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.612093 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-config\") pod \"0f2de188-5276-445c-a166-c4ffd0a68380\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.612131 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-dns-svc\") pod \"0f2de188-5276-445c-a166-c4ffd0a68380\" (UID: \"0f2de188-5276-445c-a166-c4ffd0a68380\") " Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.680150 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "0f2de188-5276-445c-a166-c4ffd0a68380" (UID: "0f2de188-5276-445c-a166-c4ffd0a68380"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.685284 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f2de188-5276-445c-a166-c4ffd0a68380-kube-api-access-7hx47" (OuterVolumeSpecName: "kube-api-access-7hx47") pod "0f2de188-5276-445c-a166-c4ffd0a68380" (UID: "0f2de188-5276-445c-a166-c4ffd0a68380"). InnerVolumeSpecName "kube-api-access-7hx47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.693585 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0f2de188-5276-445c-a166-c4ffd0a68380" (UID: "0f2de188-5276-445c-a166-c4ffd0a68380"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.700659 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0f2de188-5276-445c-a166-c4ffd0a68380" (UID: "0f2de188-5276-445c-a166-c4ffd0a68380"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.708687 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0f2de188-5276-445c-a166-c4ffd0a68380" (UID: "0f2de188-5276-445c-a166-c4ffd0a68380"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.710473 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-config" (OuterVolumeSpecName: "config") pod "0f2de188-5276-445c-a166-c4ffd0a68380" (UID: "0f2de188-5276-445c-a166-c4ffd0a68380"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.714592 4770 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.714621 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hx47\" (UniqueName: \"kubernetes.io/projected/0f2de188-5276-445c-a166-c4ffd0a68380-kube-api-access-7hx47\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.714634 4770 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.714643 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.714651 4770 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.714660 4770 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f2de188-5276-445c-a166-c4ffd0a68380-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.784466 4770 scope.go:117] "RemoveContainer" containerID="568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.946813 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64c78898cc-v6nfg"] Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.947779 4770 scope.go:117] "RemoveContainer" containerID="4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25" Oct 04 04:54:53 crc kubenswrapper[4770]: E1004 04:54:53.948712 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25\": container with ID starting with 4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25 not found: ID does not exist" containerID="4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.948748 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25"} err="failed to get container status \"4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25\": rpc error: code = NotFound desc = could not find container \"4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25\": container with ID starting with 4a14bbc608805b2817e091a8c349bd3735f9d25d390191ae15c5d3bb329e5c25 not found: ID does not exist" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.948768 4770 scope.go:117] "RemoveContainer" containerID="568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b" Oct 04 04:54:53 crc kubenswrapper[4770]: E1004 04:54:53.949075 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b\": container with ID starting with 568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b not found: ID does not exist" containerID="568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.949101 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b"} err="failed to get container status \"568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b\": rpc error: code = NotFound desc = could not find container \"568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b\": container with ID starting with 568d5dc6a71e27db131d26417070e853d74383a48a24730a398c2780f83dd50b not found: ID does not exist" Oct 04 04:54:53 crc kubenswrapper[4770]: I1004 04:54:53.957632 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64c78898cc-v6nfg"] Oct 04 04:54:55 crc kubenswrapper[4770]: I1004 04:54:55.690899 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f2de188-5276-445c-a166-c4ffd0a68380" path="/var/lib/kubelet/pods/0f2de188-5276-445c-a166-c4ffd0a68380/volumes" Oct 04 04:54:56 crc kubenswrapper[4770]: I1004 04:54:56.061717 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-mjbrc"] Oct 04 04:54:56 crc kubenswrapper[4770]: I1004 04:54:56.077427 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-mjbrc"] Oct 04 04:54:57 crc kubenswrapper[4770]: I1004 04:54:57.693794 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="109f402b-830b-4988-9a1f-f1d56652be2c" path="/var/lib/kubelet/pods/109f402b-830b-4988-9a1f-f1d56652be2c/volumes" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.867217 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx"] Oct 04 04:55:02 crc kubenswrapper[4770]: E1004 04:55:02.881584 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2de188-5276-445c-a166-c4ffd0a68380" containerName="init" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.881618 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2de188-5276-445c-a166-c4ffd0a68380" containerName="init" Oct 04 04:55:02 crc kubenswrapper[4770]: E1004 04:55:02.881682 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2de188-5276-445c-a166-c4ffd0a68380" containerName="dnsmasq-dns" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.881691 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2de188-5276-445c-a166-c4ffd0a68380" containerName="dnsmasq-dns" Oct 04 04:55:02 crc kubenswrapper[4770]: E1004 04:55:02.881735 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="529a8000-f96c-4d6e-9b47-28304e34975b" containerName="dnsmasq-dns" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.881743 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="529a8000-f96c-4d6e-9b47-28304e34975b" containerName="dnsmasq-dns" Oct 04 04:55:02 crc kubenswrapper[4770]: E1004 04:55:02.881796 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="529a8000-f96c-4d6e-9b47-28304e34975b" containerName="init" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.881804 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="529a8000-f96c-4d6e-9b47-28304e34975b" containerName="init" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.882407 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="529a8000-f96c-4d6e-9b47-28304e34975b" containerName="dnsmasq-dns" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.882444 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2de188-5276-445c-a166-c4ffd0a68380" containerName="dnsmasq-dns" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.883785 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.886227 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.890229 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.890502 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.891556 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 04:55:02 crc kubenswrapper[4770]: I1004 04:55:02.908354 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx"] Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.018320 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.018461 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.018490 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.018542 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.018857 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsshb\" (UniqueName: \"kubernetes.io/projected/94268810-ff27-479b-981f-4dd1f3b52099-kube-api-access-fsshb\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.120843 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.120912 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsshb\" (UniqueName: \"kubernetes.io/projected/94268810-ff27-479b-981f-4dd1f3b52099-kube-api-access-fsshb\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.121069 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.121187 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.121209 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.128803 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.137070 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.137242 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.140422 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.141235 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsshb\" (UniqueName: \"kubernetes.io/projected/94268810-ff27-479b-981f-4dd1f3b52099-kube-api-access-fsshb\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.210374 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:03 crc kubenswrapper[4770]: I1004 04:55:03.757681 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx"] Oct 04 04:55:03 crc kubenswrapper[4770]: W1004 04:55:03.761028 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94268810_ff27_479b_981f_4dd1f3b52099.slice/crio-991b9f69f58660404d083631f24be362b942946e1c72a3f16ad5bcb1271885ce WatchSource:0}: Error finding container 991b9f69f58660404d083631f24be362b942946e1c72a3f16ad5bcb1271885ce: Status 404 returned error can't find the container with id 991b9f69f58660404d083631f24be362b942946e1c72a3f16ad5bcb1271885ce Oct 04 04:55:04 crc kubenswrapper[4770]: I1004 04:55:04.743507 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" event={"ID":"94268810-ff27-479b-981f-4dd1f3b52099","Type":"ContainerStarted","Data":"991b9f69f58660404d083631f24be362b942946e1c72a3f16ad5bcb1271885ce"} Oct 04 04:55:08 crc kubenswrapper[4770]: I1004 04:55:08.992582 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gthzw"] Oct 04 04:55:08 crc kubenswrapper[4770]: I1004 04:55:08.995585 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.000361 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gthzw"] Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.066926 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-utilities\") pod \"certified-operators-gthzw\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.067230 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tw25\" (UniqueName: \"kubernetes.io/projected/d64699f3-18a1-45f4-b9b1-784d781851f9-kube-api-access-5tw25\") pod \"certified-operators-gthzw\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.067464 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-catalog-content\") pod \"certified-operators-gthzw\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.169940 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-utilities\") pod \"certified-operators-gthzw\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.169989 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tw25\" (UniqueName: \"kubernetes.io/projected/d64699f3-18a1-45f4-b9b1-784d781851f9-kube-api-access-5tw25\") pod \"certified-operators-gthzw\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.170189 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-catalog-content\") pod \"certified-operators-gthzw\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.171123 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-utilities\") pod \"certified-operators-gthzw\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.172295 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-catalog-content\") pod \"certified-operators-gthzw\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.199912 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tw25\" (UniqueName: \"kubernetes.io/projected/d64699f3-18a1-45f4-b9b1-784d781851f9-kube-api-access-5tw25\") pod \"certified-operators-gthzw\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:09 crc kubenswrapper[4770]: I1004 04:55:09.332085 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:16 crc kubenswrapper[4770]: E1004 04:55:16.854340 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Oct 04 04:55:16 crc kubenswrapper[4770]: E1004 04:55:16.855065 4770 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 04 04:55:16 crc kubenswrapper[4770]: container &Container{Name:pre-adoption-validation-openstack-pre-adoption-openstack-cell1,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p osp.edpm.pre_adoption_validation -i pre-adoption-validation-openstack-pre-adoption-openstack-cell1],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_CALLBACKS_ENABLED,Value:ansible.posix.profile_tasks,yaml,ValueFrom:nil,},EnvVar{Name:ANSIBLE_FORCE_COLOR,Value:True,ValueFrom:nil,},EnvVar{Name:ANSIBLE_DISPLAY_ARGS_TO_STDOUT,Value:True,ValueFrom:nil,},EnvVar{Name:ANSIBLE_SSH_ARGS,Value:-C -o ControlMaster=auto -o ControlPersist=80s,ValueFrom:nil,},EnvVar{Name:ANSIBLE_VERBOSITY,Value:1,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Oct 04 04:55:16 crc kubenswrapper[4770]: osp.edpm.pre_adoption_validation Oct 04 04:55:16 crc kubenswrapper[4770]: Oct 04 04:55:16 crc kubenswrapper[4770]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Oct 04 04:55:16 crc kubenswrapper[4770]: edpm_override_hosts: openstack-cell1 Oct 04 04:55:16 crc kubenswrapper[4770]: edpm_service_type: pre-adoption-validation Oct 04 04:55:16 crc kubenswrapper[4770]: edpm_services_override: [pre-adoption-validation] Oct 04 04:55:16 crc kubenswrapper[4770]: Oct 04 04:55:16 crc kubenswrapper[4770]: Oct 04 04:55:16 crc kubenswrapper[4770]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:pre-adoption-validation-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/pre-adoption-validation,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fsshb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx_openstack(94268810-ff27-479b-981f-4dd1f3b52099): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Oct 04 04:55:16 crc kubenswrapper[4770]: > logger="UnhandledError" Oct 04 04:55:16 crc kubenswrapper[4770]: E1004 04:55:16.856291 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"pre-adoption-validation-openstack-pre-adoption-openstack-cell1\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" podUID="94268810-ff27-479b-981f-4dd1f3b52099" Oct 04 04:55:16 crc kubenswrapper[4770]: I1004 04:55:16.860998 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gthzw"] Oct 04 04:55:16 crc kubenswrapper[4770]: I1004 04:55:16.922348 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gthzw" event={"ID":"d64699f3-18a1-45f4-b9b1-784d781851f9","Type":"ContainerStarted","Data":"d6a3d4ac469bcc1c8ac95d6669d0f2fa2ca273193d0488f3aceb1193a4a59f61"} Oct 04 04:55:16 crc kubenswrapper[4770]: E1004 04:55:16.923843 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"pre-adoption-validation-openstack-pre-adoption-openstack-cell1\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" podUID="94268810-ff27-479b-981f-4dd1f3b52099" Oct 04 04:55:17 crc kubenswrapper[4770]: I1004 04:55:17.931908 4770 generic.go:334] "Generic (PLEG): container finished" podID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerID="bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc" exitCode=0 Oct 04 04:55:17 crc kubenswrapper[4770]: I1004 04:55:17.931969 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gthzw" event={"ID":"d64699f3-18a1-45f4-b9b1-784d781851f9","Type":"ContainerDied","Data":"bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc"} Oct 04 04:55:19 crc kubenswrapper[4770]: I1004 04:55:19.955241 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gthzw" event={"ID":"d64699f3-18a1-45f4-b9b1-784d781851f9","Type":"ContainerStarted","Data":"432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499"} Oct 04 04:55:24 crc kubenswrapper[4770]: I1004 04:55:24.009123 4770 generic.go:334] "Generic (PLEG): container finished" podID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerID="432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499" exitCode=0 Oct 04 04:55:24 crc kubenswrapper[4770]: I1004 04:55:24.009199 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gthzw" event={"ID":"d64699f3-18a1-45f4-b9b1-784d781851f9","Type":"ContainerDied","Data":"432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499"} Oct 04 04:55:29 crc kubenswrapper[4770]: I1004 04:55:29.948286 4770 scope.go:117] "RemoveContainer" containerID="685520aeef311df4260e8f8767f9d6baef1790b2fc015bca3bcbb729e1f612ad" Oct 04 04:55:30 crc kubenswrapper[4770]: I1004 04:55:30.006288 4770 scope.go:117] "RemoveContainer" containerID="86a253a749e1143c875200c889641eaf602c948aea1da60992c163c8c90b94af" Oct 04 04:55:31 crc kubenswrapper[4770]: I1004 04:55:31.796457 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:55:31 crc kubenswrapper[4770]: I1004 04:55:31.796804 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:55:37 crc kubenswrapper[4770]: I1004 04:55:37.061889 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:55:38 crc kubenswrapper[4770]: I1004 04:55:38.157581 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gthzw" event={"ID":"d64699f3-18a1-45f4-b9b1-784d781851f9","Type":"ContainerStarted","Data":"c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c"} Oct 04 04:55:38 crc kubenswrapper[4770]: I1004 04:55:38.159990 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" event={"ID":"94268810-ff27-479b-981f-4dd1f3b52099","Type":"ContainerStarted","Data":"3b0ab8c31fd4dc52ef4e127533548fd4816593056858da22fd5cdf1a371c7fa8"} Oct 04 04:55:38 crc kubenswrapper[4770]: I1004 04:55:38.181889 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gthzw" podStartSLOduration=11.079008367 podStartE2EDuration="30.18187343s" podCreationTimestamp="2025-10-04 04:55:08 +0000 UTC" firstStartedPulling="2025-10-04 04:55:17.93362903 +0000 UTC m=+6729.225638762" lastFinishedPulling="2025-10-04 04:55:37.036494113 +0000 UTC m=+6748.328503825" observedRunningTime="2025-10-04 04:55:38.175381861 +0000 UTC m=+6749.467391583" watchObservedRunningTime="2025-10-04 04:55:38.18187343 +0000 UTC m=+6749.473883142" Oct 04 04:55:38 crc kubenswrapper[4770]: I1004 04:55:38.206549 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" podStartSLOduration=2.911398439 podStartE2EDuration="36.206527895s" podCreationTimestamp="2025-10-04 04:55:02 +0000 UTC" firstStartedPulling="2025-10-04 04:55:03.764676647 +0000 UTC m=+6715.056686359" lastFinishedPulling="2025-10-04 04:55:37.059806083 +0000 UTC m=+6748.351815815" observedRunningTime="2025-10-04 04:55:38.197696275 +0000 UTC m=+6749.489706007" watchObservedRunningTime="2025-10-04 04:55:38.206527895 +0000 UTC m=+6749.498537617" Oct 04 04:55:39 crc kubenswrapper[4770]: I1004 04:55:39.332562 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:39 crc kubenswrapper[4770]: I1004 04:55:39.333106 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:40 crc kubenswrapper[4770]: I1004 04:55:40.399852 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-gthzw" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerName="registry-server" probeResult="failure" output=< Oct 04 04:55:40 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 04:55:40 crc kubenswrapper[4770]: > Oct 04 04:55:49 crc kubenswrapper[4770]: I1004 04:55:49.400260 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:49 crc kubenswrapper[4770]: I1004 04:55:49.471599 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:49 crc kubenswrapper[4770]: I1004 04:55:49.666514 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gthzw"] Oct 04 04:55:51 crc kubenswrapper[4770]: I1004 04:55:51.323084 4770 generic.go:334] "Generic (PLEG): container finished" podID="94268810-ff27-479b-981f-4dd1f3b52099" containerID="3b0ab8c31fd4dc52ef4e127533548fd4816593056858da22fd5cdf1a371c7fa8" exitCode=0 Oct 04 04:55:51 crc kubenswrapper[4770]: I1004 04:55:51.323134 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" event={"ID":"94268810-ff27-479b-981f-4dd1f3b52099","Type":"ContainerDied","Data":"3b0ab8c31fd4dc52ef4e127533548fd4816593056858da22fd5cdf1a371c7fa8"} Oct 04 04:55:51 crc kubenswrapper[4770]: I1004 04:55:51.323617 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gthzw" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerName="registry-server" containerID="cri-o://c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c" gracePeriod=2 Oct 04 04:55:51 crc kubenswrapper[4770]: I1004 04:55:51.864767 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.004791 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-catalog-content\") pod \"d64699f3-18a1-45f4-b9b1-784d781851f9\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.005147 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tw25\" (UniqueName: \"kubernetes.io/projected/d64699f3-18a1-45f4-b9b1-784d781851f9-kube-api-access-5tw25\") pod \"d64699f3-18a1-45f4-b9b1-784d781851f9\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.005230 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-utilities\") pod \"d64699f3-18a1-45f4-b9b1-784d781851f9\" (UID: \"d64699f3-18a1-45f4-b9b1-784d781851f9\") " Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.006440 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-utilities" (OuterVolumeSpecName: "utilities") pod "d64699f3-18a1-45f4-b9b1-784d781851f9" (UID: "d64699f3-18a1-45f4-b9b1-784d781851f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.017366 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64699f3-18a1-45f4-b9b1-784d781851f9-kube-api-access-5tw25" (OuterVolumeSpecName: "kube-api-access-5tw25") pod "d64699f3-18a1-45f4-b9b1-784d781851f9" (UID: "d64699f3-18a1-45f4-b9b1-784d781851f9"). InnerVolumeSpecName "kube-api-access-5tw25". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.085415 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d64699f3-18a1-45f4-b9b1-784d781851f9" (UID: "d64699f3-18a1-45f4-b9b1-784d781851f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.107468 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tw25\" (UniqueName: \"kubernetes.io/projected/d64699f3-18a1-45f4-b9b1-784d781851f9-kube-api-access-5tw25\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.107501 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.107511 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64699f3-18a1-45f4-b9b1-784d781851f9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.335310 4770 generic.go:334] "Generic (PLEG): container finished" podID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerID="c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c" exitCode=0 Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.335390 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gthzw" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.335467 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gthzw" event={"ID":"d64699f3-18a1-45f4-b9b1-784d781851f9","Type":"ContainerDied","Data":"c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c"} Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.335507 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gthzw" event={"ID":"d64699f3-18a1-45f4-b9b1-784d781851f9","Type":"ContainerDied","Data":"d6a3d4ac469bcc1c8ac95d6669d0f2fa2ca273193d0488f3aceb1193a4a59f61"} Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.335530 4770 scope.go:117] "RemoveContainer" containerID="c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.375501 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gthzw"] Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.385612 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gthzw"] Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.386828 4770 scope.go:117] "RemoveContainer" containerID="432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.434729 4770 scope.go:117] "RemoveContainer" containerID="bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.495567 4770 scope.go:117] "RemoveContainer" containerID="c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c" Oct 04 04:55:52 crc kubenswrapper[4770]: E1004 04:55:52.496216 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c\": container with ID starting with c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c not found: ID does not exist" containerID="c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.496264 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c"} err="failed to get container status \"c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c\": rpc error: code = NotFound desc = could not find container \"c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c\": container with ID starting with c64a6fd8863bd03249f74f185379720ff534ee2207ac9fe254f2c5616356944c not found: ID does not exist" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.496297 4770 scope.go:117] "RemoveContainer" containerID="432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499" Oct 04 04:55:52 crc kubenswrapper[4770]: E1004 04:55:52.497917 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499\": container with ID starting with 432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499 not found: ID does not exist" containerID="432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.497941 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499"} err="failed to get container status \"432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499\": rpc error: code = NotFound desc = could not find container \"432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499\": container with ID starting with 432317c7450fa460e1b218dd7a561cc79d02f7121eeae455410d942092561499 not found: ID does not exist" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.497957 4770 scope.go:117] "RemoveContainer" containerID="bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc" Oct 04 04:55:52 crc kubenswrapper[4770]: E1004 04:55:52.498219 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc\": container with ID starting with bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc not found: ID does not exist" containerID="bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.498242 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc"} err="failed to get container status \"bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc\": rpc error: code = NotFound desc = could not find container \"bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc\": container with ID starting with bad1a36aba74abad3568c202fa2b386c498e462944cd13161b045544626690fc not found: ID does not exist" Oct 04 04:55:52 crc kubenswrapper[4770]: I1004 04:55:52.860215 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.025244 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ssh-key\") pod \"94268810-ff27-479b-981f-4dd1f3b52099\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.025313 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-inventory\") pod \"94268810-ff27-479b-981f-4dd1f3b52099\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.025420 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ceph\") pod \"94268810-ff27-479b-981f-4dd1f3b52099\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.025541 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsshb\" (UniqueName: \"kubernetes.io/projected/94268810-ff27-479b-981f-4dd1f3b52099-kube-api-access-fsshb\") pod \"94268810-ff27-479b-981f-4dd1f3b52099\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.025663 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-pre-adoption-validation-combined-ca-bundle\") pod \"94268810-ff27-479b-981f-4dd1f3b52099\" (UID: \"94268810-ff27-479b-981f-4dd1f3b52099\") " Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.031164 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "94268810-ff27-479b-981f-4dd1f3b52099" (UID: "94268810-ff27-479b-981f-4dd1f3b52099"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.040394 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94268810-ff27-479b-981f-4dd1f3b52099-kube-api-access-fsshb" (OuterVolumeSpecName: "kube-api-access-fsshb") pod "94268810-ff27-479b-981f-4dd1f3b52099" (UID: "94268810-ff27-479b-981f-4dd1f3b52099"). InnerVolumeSpecName "kube-api-access-fsshb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.041969 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ceph" (OuterVolumeSpecName: "ceph") pod "94268810-ff27-479b-981f-4dd1f3b52099" (UID: "94268810-ff27-479b-981f-4dd1f3b52099"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.060146 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "94268810-ff27-479b-981f-4dd1f3b52099" (UID: "94268810-ff27-479b-981f-4dd1f3b52099"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.078168 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-inventory" (OuterVolumeSpecName: "inventory") pod "94268810-ff27-479b-981f-4dd1f3b52099" (UID: "94268810-ff27-479b-981f-4dd1f3b52099"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.129598 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsshb\" (UniqueName: \"kubernetes.io/projected/94268810-ff27-479b-981f-4dd1f3b52099-kube-api-access-fsshb\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.129636 4770 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.129648 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.129657 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.129667 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/94268810-ff27-479b-981f-4dd1f3b52099-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.350308 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" event={"ID":"94268810-ff27-479b-981f-4dd1f3b52099","Type":"ContainerDied","Data":"991b9f69f58660404d083631f24be362b942946e1c72a3f16ad5bcb1271885ce"} Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.350338 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.350369 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="991b9f69f58660404d083631f24be362b942946e1c72a3f16ad5bcb1271885ce" Oct 04 04:55:53 crc kubenswrapper[4770]: I1004 04:55:53.684773 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" path="/var/lib/kubelet/pods/d64699f3-18a1-45f4-b9b1-784d781851f9/volumes" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.339845 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848"] Oct 04 04:55:56 crc kubenswrapper[4770]: E1004 04:55:56.340747 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerName="extract-utilities" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.340759 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerName="extract-utilities" Oct 04 04:55:56 crc kubenswrapper[4770]: E1004 04:55:56.340782 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerName="extract-content" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.340787 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerName="extract-content" Oct 04 04:55:56 crc kubenswrapper[4770]: E1004 04:55:56.340816 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerName="registry-server" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.340822 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerName="registry-server" Oct 04 04:55:56 crc kubenswrapper[4770]: E1004 04:55:56.340833 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94268810-ff27-479b-981f-4dd1f3b52099" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.340842 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="94268810-ff27-479b-981f-4dd1f3b52099" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.341061 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="94268810-ff27-479b-981f-4dd1f3b52099" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.341095 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64699f3-18a1-45f4-b9b1-784d781851f9" containerName="registry-server" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.341891 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.350496 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.350559 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.350973 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.351250 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.373118 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848"] Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.511508 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgbfd\" (UniqueName: \"kubernetes.io/projected/b5dcbbe1-7d4d-4c82-a705-f367cb129883-kube-api-access-rgbfd\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.511598 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.511647 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.511762 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.511786 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.613472 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.613599 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.613847 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.613893 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.613995 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgbfd\" (UniqueName: \"kubernetes.io/projected/b5dcbbe1-7d4d-4c82-a705-f367cb129883-kube-api-access-rgbfd\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.621927 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.621938 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.622126 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.622298 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.636971 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgbfd\" (UniqueName: \"kubernetes.io/projected/b5dcbbe1-7d4d-4c82-a705-f367cb129883-kube-api-access-rgbfd\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:56 crc kubenswrapper[4770]: I1004 04:55:56.668400 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 04:55:57 crc kubenswrapper[4770]: I1004 04:55:57.232703 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848"] Oct 04 04:55:57 crc kubenswrapper[4770]: I1004 04:55:57.396831 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" event={"ID":"b5dcbbe1-7d4d-4c82-a705-f367cb129883","Type":"ContainerStarted","Data":"bd5ef5e5325343ea89d4c4a571656be2a461cc1f710efa4b03a27f02b8bdd1e9"} Oct 04 04:55:59 crc kubenswrapper[4770]: I1004 04:55:59.425268 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" event={"ID":"b5dcbbe1-7d4d-4c82-a705-f367cb129883","Type":"ContainerStarted","Data":"2250a68a23d625dd77b74e9e040ec40210433d1a04fafcd3c2da6f9eb2992c39"} Oct 04 04:55:59 crc kubenswrapper[4770]: I1004 04:55:59.448789 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" podStartSLOduration=2.3792279880000002 podStartE2EDuration="3.448771122s" podCreationTimestamp="2025-10-04 04:55:56 +0000 UTC" firstStartedPulling="2025-10-04 04:55:57.23918893 +0000 UTC m=+6768.531198642" lastFinishedPulling="2025-10-04 04:55:58.308732054 +0000 UTC m=+6769.600741776" observedRunningTime="2025-10-04 04:55:59.443125074 +0000 UTC m=+6770.735134826" watchObservedRunningTime="2025-10-04 04:55:59.448771122 +0000 UTC m=+6770.740780834" Oct 04 04:56:01 crc kubenswrapper[4770]: I1004 04:56:01.795958 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:56:01 crc kubenswrapper[4770]: I1004 04:56:01.796345 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:56:31 crc kubenswrapper[4770]: I1004 04:56:31.796095 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:56:31 crc kubenswrapper[4770]: I1004 04:56:31.797098 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:56:31 crc kubenswrapper[4770]: I1004 04:56:31.797175 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 04:56:31 crc kubenswrapper[4770]: I1004 04:56:31.798500 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:56:31 crc kubenswrapper[4770]: I1004 04:56:31.798580 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" gracePeriod=600 Oct 04 04:56:31 crc kubenswrapper[4770]: E1004 04:56:31.948115 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:56:32 crc kubenswrapper[4770]: I1004 04:56:32.781808 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" exitCode=0 Oct 04 04:56:32 crc kubenswrapper[4770]: I1004 04:56:32.782054 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac"} Oct 04 04:56:32 crc kubenswrapper[4770]: I1004 04:56:32.782321 4770 scope.go:117] "RemoveContainer" containerID="284e8fec1a55dcb36285518859a2a4bc58d3161ca3219d7437f89da44e6444b6" Oct 04 04:56:32 crc kubenswrapper[4770]: I1004 04:56:32.783223 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:56:32 crc kubenswrapper[4770]: E1004 04:56:32.783615 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:56:47 crc kubenswrapper[4770]: I1004 04:56:47.674176 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:56:47 crc kubenswrapper[4770]: E1004 04:56:47.675125 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:56:59 crc kubenswrapper[4770]: I1004 04:56:59.681267 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:56:59 crc kubenswrapper[4770]: E1004 04:56:59.681883 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:57:11 crc kubenswrapper[4770]: I1004 04:57:11.674455 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:57:11 crc kubenswrapper[4770]: E1004 04:57:11.675424 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:57:22 crc kubenswrapper[4770]: I1004 04:57:22.673898 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:57:22 crc kubenswrapper[4770]: E1004 04:57:22.674827 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:57:37 crc kubenswrapper[4770]: I1004 04:57:37.674741 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:57:37 crc kubenswrapper[4770]: E1004 04:57:37.676384 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:57:48 crc kubenswrapper[4770]: I1004 04:57:48.674101 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:57:48 crc kubenswrapper[4770]: E1004 04:57:48.674944 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:57:59 crc kubenswrapper[4770]: I1004 04:57:59.681902 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:57:59 crc kubenswrapper[4770]: E1004 04:57:59.682831 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:58:02 crc kubenswrapper[4770]: I1004 04:58:02.837910 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l4rnc"] Oct 04 04:58:02 crc kubenswrapper[4770]: I1004 04:58:02.842711 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:02 crc kubenswrapper[4770]: I1004 04:58:02.850967 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4rnc"] Oct 04 04:58:02 crc kubenswrapper[4770]: I1004 04:58:02.994431 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-utilities\") pod \"redhat-marketplace-l4rnc\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:02 crc kubenswrapper[4770]: I1004 04:58:02.994530 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-catalog-content\") pod \"redhat-marketplace-l4rnc\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:02 crc kubenswrapper[4770]: I1004 04:58:02.994716 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmq4v\" (UniqueName: \"kubernetes.io/projected/9971ba2a-356e-4c80-ad18-52f639687887-kube-api-access-mmq4v\") pod \"redhat-marketplace-l4rnc\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:03 crc kubenswrapper[4770]: I1004 04:58:03.096816 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-catalog-content\") pod \"redhat-marketplace-l4rnc\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:03 crc kubenswrapper[4770]: I1004 04:58:03.097021 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmq4v\" (UniqueName: \"kubernetes.io/projected/9971ba2a-356e-4c80-ad18-52f639687887-kube-api-access-mmq4v\") pod \"redhat-marketplace-l4rnc\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:03 crc kubenswrapper[4770]: I1004 04:58:03.097246 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-utilities\") pod \"redhat-marketplace-l4rnc\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:03 crc kubenswrapper[4770]: I1004 04:58:03.097506 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-catalog-content\") pod \"redhat-marketplace-l4rnc\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:03 crc kubenswrapper[4770]: I1004 04:58:03.097936 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-utilities\") pod \"redhat-marketplace-l4rnc\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:03 crc kubenswrapper[4770]: I1004 04:58:03.121977 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmq4v\" (UniqueName: \"kubernetes.io/projected/9971ba2a-356e-4c80-ad18-52f639687887-kube-api-access-mmq4v\") pod \"redhat-marketplace-l4rnc\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:03 crc kubenswrapper[4770]: I1004 04:58:03.169415 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:03 crc kubenswrapper[4770]: I1004 04:58:03.624586 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4rnc"] Oct 04 04:58:03 crc kubenswrapper[4770]: I1004 04:58:03.808464 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4rnc" event={"ID":"9971ba2a-356e-4c80-ad18-52f639687887","Type":"ContainerStarted","Data":"0a768a6a9485e0723fc14e23cd08641dd920494cbed94ecec88994ede7bda0c8"} Oct 04 04:58:04 crc kubenswrapper[4770]: I1004 04:58:04.819717 4770 generic.go:334] "Generic (PLEG): container finished" podID="9971ba2a-356e-4c80-ad18-52f639687887" containerID="325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba" exitCode=0 Oct 04 04:58:04 crc kubenswrapper[4770]: I1004 04:58:04.819769 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4rnc" event={"ID":"9971ba2a-356e-4c80-ad18-52f639687887","Type":"ContainerDied","Data":"325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba"} Oct 04 04:58:06 crc kubenswrapper[4770]: I1004 04:58:06.843705 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4rnc" event={"ID":"9971ba2a-356e-4c80-ad18-52f639687887","Type":"ContainerStarted","Data":"2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6"} Oct 04 04:58:07 crc kubenswrapper[4770]: I1004 04:58:07.872526 4770 generic.go:334] "Generic (PLEG): container finished" podID="9971ba2a-356e-4c80-ad18-52f639687887" containerID="2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6" exitCode=0 Oct 04 04:58:07 crc kubenswrapper[4770]: I1004 04:58:07.872570 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4rnc" event={"ID":"9971ba2a-356e-4c80-ad18-52f639687887","Type":"ContainerDied","Data":"2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6"} Oct 04 04:58:09 crc kubenswrapper[4770]: I1004 04:58:09.900038 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4rnc" event={"ID":"9971ba2a-356e-4c80-ad18-52f639687887","Type":"ContainerStarted","Data":"bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e"} Oct 04 04:58:09 crc kubenswrapper[4770]: I1004 04:58:09.928281 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l4rnc" podStartSLOduration=3.366928092 podStartE2EDuration="7.928260645s" podCreationTimestamp="2025-10-04 04:58:02 +0000 UTC" firstStartedPulling="2025-10-04 04:58:04.824629728 +0000 UTC m=+6896.116639440" lastFinishedPulling="2025-10-04 04:58:09.385962291 +0000 UTC m=+6900.677971993" observedRunningTime="2025-10-04 04:58:09.917274368 +0000 UTC m=+6901.209284080" watchObservedRunningTime="2025-10-04 04:58:09.928260645 +0000 UTC m=+6901.220270357" Oct 04 04:58:10 crc kubenswrapper[4770]: I1004 04:58:10.673740 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:58:10 crc kubenswrapper[4770]: E1004 04:58:10.674333 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:58:13 crc kubenswrapper[4770]: I1004 04:58:13.170032 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:13 crc kubenswrapper[4770]: I1004 04:58:13.171267 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:13 crc kubenswrapper[4770]: I1004 04:58:13.240689 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:23 crc kubenswrapper[4770]: I1004 04:58:23.235637 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:23 crc kubenswrapper[4770]: I1004 04:58:23.286668 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4rnc"] Oct 04 04:58:23 crc kubenswrapper[4770]: I1004 04:58:23.673611 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:58:23 crc kubenswrapper[4770]: E1004 04:58:23.673874 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.080957 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l4rnc" podUID="9971ba2a-356e-4c80-ad18-52f639687887" containerName="registry-server" containerID="cri-o://bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e" gracePeriod=2 Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.643206 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.789698 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmq4v\" (UniqueName: \"kubernetes.io/projected/9971ba2a-356e-4c80-ad18-52f639687887-kube-api-access-mmq4v\") pod \"9971ba2a-356e-4c80-ad18-52f639687887\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.789898 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-utilities\") pod \"9971ba2a-356e-4c80-ad18-52f639687887\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.789969 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-catalog-content\") pod \"9971ba2a-356e-4c80-ad18-52f639687887\" (UID: \"9971ba2a-356e-4c80-ad18-52f639687887\") " Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.790632 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-utilities" (OuterVolumeSpecName: "utilities") pod "9971ba2a-356e-4c80-ad18-52f639687887" (UID: "9971ba2a-356e-4c80-ad18-52f639687887"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.790737 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.795083 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9971ba2a-356e-4c80-ad18-52f639687887-kube-api-access-mmq4v" (OuterVolumeSpecName: "kube-api-access-mmq4v") pod "9971ba2a-356e-4c80-ad18-52f639687887" (UID: "9971ba2a-356e-4c80-ad18-52f639687887"). InnerVolumeSpecName "kube-api-access-mmq4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.804054 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9971ba2a-356e-4c80-ad18-52f639687887" (UID: "9971ba2a-356e-4c80-ad18-52f639687887"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.893259 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9971ba2a-356e-4c80-ad18-52f639687887-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:24 crc kubenswrapper[4770]: I1004 04:58:24.893771 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmq4v\" (UniqueName: \"kubernetes.io/projected/9971ba2a-356e-4c80-ad18-52f639687887-kube-api-access-mmq4v\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.092048 4770 generic.go:334] "Generic (PLEG): container finished" podID="9971ba2a-356e-4c80-ad18-52f639687887" containerID="bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e" exitCode=0 Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.092135 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4rnc" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.092135 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4rnc" event={"ID":"9971ba2a-356e-4c80-ad18-52f639687887","Type":"ContainerDied","Data":"bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e"} Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.092208 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4rnc" event={"ID":"9971ba2a-356e-4c80-ad18-52f639687887","Type":"ContainerDied","Data":"0a768a6a9485e0723fc14e23cd08641dd920494cbed94ecec88994ede7bda0c8"} Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.092237 4770 scope.go:117] "RemoveContainer" containerID="bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.126380 4770 scope.go:117] "RemoveContainer" containerID="2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.126840 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4rnc"] Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.137363 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4rnc"] Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.148835 4770 scope.go:117] "RemoveContainer" containerID="325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.219738 4770 scope.go:117] "RemoveContainer" containerID="bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e" Oct 04 04:58:25 crc kubenswrapper[4770]: E1004 04:58:25.220570 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e\": container with ID starting with bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e not found: ID does not exist" containerID="bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.220608 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e"} err="failed to get container status \"bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e\": rpc error: code = NotFound desc = could not find container \"bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e\": container with ID starting with bc88e3d03eff2517b4cc095c6bc1db41dca263f51bfba610d896263f57de9d7e not found: ID does not exist" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.220633 4770 scope.go:117] "RemoveContainer" containerID="2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6" Oct 04 04:58:25 crc kubenswrapper[4770]: E1004 04:58:25.220991 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6\": container with ID starting with 2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6 not found: ID does not exist" containerID="2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.221030 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6"} err="failed to get container status \"2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6\": rpc error: code = NotFound desc = could not find container \"2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6\": container with ID starting with 2d029915ba814d888354e840bd71c6185efe9c50be7efd201fab328e40dce6a6 not found: ID does not exist" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.221051 4770 scope.go:117] "RemoveContainer" containerID="325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba" Oct 04 04:58:25 crc kubenswrapper[4770]: E1004 04:58:25.221280 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba\": container with ID starting with 325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba not found: ID does not exist" containerID="325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.221305 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba"} err="failed to get container status \"325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba\": rpc error: code = NotFound desc = could not find container \"325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba\": container with ID starting with 325df42c9856d09bcc87d66c6e6f120034d690e27c632edb6a4818d3978a71ba not found: ID does not exist" Oct 04 04:58:25 crc kubenswrapper[4770]: I1004 04:58:25.691105 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9971ba2a-356e-4c80-ad18-52f639687887" path="/var/lib/kubelet/pods/9971ba2a-356e-4c80-ad18-52f639687887/volumes" Oct 04 04:58:35 crc kubenswrapper[4770]: I1004 04:58:35.674247 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:58:35 crc kubenswrapper[4770]: E1004 04:58:35.675193 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:58:41 crc kubenswrapper[4770]: I1004 04:58:41.058205 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-c9z68"] Oct 04 04:58:41 crc kubenswrapper[4770]: I1004 04:58:41.071383 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-c9z68"] Oct 04 04:58:41 crc kubenswrapper[4770]: I1004 04:58:41.686141 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb09598e-d547-4375-a495-0f14fe0d0480" path="/var/lib/kubelet/pods/bb09598e-d547-4375-a495-0f14fe0d0480/volumes" Oct 04 04:58:47 crc kubenswrapper[4770]: I1004 04:58:47.674395 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:58:47 crc kubenswrapper[4770]: E1004 04:58:47.675247 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:58:50 crc kubenswrapper[4770]: I1004 04:58:50.065602 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-574f-account-create-2vw9g"] Oct 04 04:58:50 crc kubenswrapper[4770]: I1004 04:58:50.078297 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-574f-account-create-2vw9g"] Oct 04 04:58:51 crc kubenswrapper[4770]: I1004 04:58:51.687349 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef122e98-17e7-49d3-9e07-24cf340332b5" path="/var/lib/kubelet/pods/ef122e98-17e7-49d3-9e07-24cf340332b5/volumes" Oct 04 04:59:01 crc kubenswrapper[4770]: I1004 04:59:01.674929 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:59:01 crc kubenswrapper[4770]: E1004 04:59:01.676467 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:59:07 crc kubenswrapper[4770]: I1004 04:59:07.075041 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-fjp9q"] Oct 04 04:59:07 crc kubenswrapper[4770]: I1004 04:59:07.084517 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-fjp9q"] Oct 04 04:59:07 crc kubenswrapper[4770]: I1004 04:59:07.688193 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="514cf20d-92df-4de4-a2b7-5607902ba081" path="/var/lib/kubelet/pods/514cf20d-92df-4de4-a2b7-5607902ba081/volumes" Oct 04 04:59:14 crc kubenswrapper[4770]: I1004 04:59:14.674759 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:59:14 crc kubenswrapper[4770]: E1004 04:59:14.675586 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.398254 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tp8f7"] Oct 04 04:59:26 crc kubenswrapper[4770]: E1004 04:59:26.399294 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9971ba2a-356e-4c80-ad18-52f639687887" containerName="registry-server" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.399310 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9971ba2a-356e-4c80-ad18-52f639687887" containerName="registry-server" Oct 04 04:59:26 crc kubenswrapper[4770]: E1004 04:59:26.399337 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9971ba2a-356e-4c80-ad18-52f639687887" containerName="extract-content" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.399343 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9971ba2a-356e-4c80-ad18-52f639687887" containerName="extract-content" Oct 04 04:59:26 crc kubenswrapper[4770]: E1004 04:59:26.399368 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9971ba2a-356e-4c80-ad18-52f639687887" containerName="extract-utilities" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.399374 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9971ba2a-356e-4c80-ad18-52f639687887" containerName="extract-utilities" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.399566 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9971ba2a-356e-4c80-ad18-52f639687887" containerName="registry-server" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.401175 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.410198 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tp8f7"] Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.459769 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-utilities\") pod \"redhat-operators-tp8f7\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.460208 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnj2q\" (UniqueName: \"kubernetes.io/projected/e95792e6-508b-4cb8-bc3e-35c8825c9761-kube-api-access-pnj2q\") pod \"redhat-operators-tp8f7\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.460395 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-catalog-content\") pod \"redhat-operators-tp8f7\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.562531 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnj2q\" (UniqueName: \"kubernetes.io/projected/e95792e6-508b-4cb8-bc3e-35c8825c9761-kube-api-access-pnj2q\") pod \"redhat-operators-tp8f7\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.562665 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-catalog-content\") pod \"redhat-operators-tp8f7\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.562717 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-utilities\") pod \"redhat-operators-tp8f7\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.563305 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-catalog-content\") pod \"redhat-operators-tp8f7\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.563370 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-utilities\") pod \"redhat-operators-tp8f7\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.581992 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnj2q\" (UniqueName: \"kubernetes.io/projected/e95792e6-508b-4cb8-bc3e-35c8825c9761-kube-api-access-pnj2q\") pod \"redhat-operators-tp8f7\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:26 crc kubenswrapper[4770]: I1004 04:59:26.731631 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:27 crc kubenswrapper[4770]: I1004 04:59:27.282223 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tp8f7"] Oct 04 04:59:27 crc kubenswrapper[4770]: I1004 04:59:27.830097 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tp8f7" event={"ID":"e95792e6-508b-4cb8-bc3e-35c8825c9761","Type":"ContainerStarted","Data":"edd3dc98200ed8fa0c6c9287b373c8d4167ea1c8f421a607c46a7adb09f11cee"} Oct 04 04:59:28 crc kubenswrapper[4770]: I1004 04:59:28.840231 4770 generic.go:334] "Generic (PLEG): container finished" podID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerID="aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe" exitCode=0 Oct 04 04:59:28 crc kubenswrapper[4770]: I1004 04:59:28.840280 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tp8f7" event={"ID":"e95792e6-508b-4cb8-bc3e-35c8825c9761","Type":"ContainerDied","Data":"aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe"} Oct 04 04:59:28 crc kubenswrapper[4770]: I1004 04:59:28.842619 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:59:28 crc kubenswrapper[4770]: I1004 04:59:28.994668 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wgg4m"] Oct 04 04:59:28 crc kubenswrapper[4770]: I1004 04:59:28.997444 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.006752 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wgg4m"] Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.017857 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fqsk\" (UniqueName: \"kubernetes.io/projected/38911d7b-b75c-40bf-8ab0-296e1324e78f-kube-api-access-7fqsk\") pod \"community-operators-wgg4m\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.018025 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-utilities\") pod \"community-operators-wgg4m\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.018102 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-catalog-content\") pod \"community-operators-wgg4m\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.120232 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fqsk\" (UniqueName: \"kubernetes.io/projected/38911d7b-b75c-40bf-8ab0-296e1324e78f-kube-api-access-7fqsk\") pod \"community-operators-wgg4m\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.120336 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-utilities\") pod \"community-operators-wgg4m\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.120406 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-catalog-content\") pod \"community-operators-wgg4m\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.120820 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-catalog-content\") pod \"community-operators-wgg4m\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.120817 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-utilities\") pod \"community-operators-wgg4m\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.142804 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fqsk\" (UniqueName: \"kubernetes.io/projected/38911d7b-b75c-40bf-8ab0-296e1324e78f-kube-api-access-7fqsk\") pod \"community-operators-wgg4m\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.320940 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.681271 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:59:29 crc kubenswrapper[4770]: E1004 04:59:29.683907 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.786560 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wgg4m"] Oct 04 04:59:29 crc kubenswrapper[4770]: I1004 04:59:29.874737 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wgg4m" event={"ID":"38911d7b-b75c-40bf-8ab0-296e1324e78f","Type":"ContainerStarted","Data":"c424ce5c391dac115c36ed90f20f7d6c4ba072134d539f113385aa4d4558daf8"} Oct 04 04:59:30 crc kubenswrapper[4770]: I1004 04:59:30.273809 4770 scope.go:117] "RemoveContainer" containerID="f1bcb7e4fe0bfa6f8ffc1cdc136c15c6fc3c41c7a30d8f60281db6ac74f086de" Oct 04 04:59:30 crc kubenswrapper[4770]: I1004 04:59:30.303261 4770 scope.go:117] "RemoveContainer" containerID="bf97667caafea1cfca3bd6c0facac44e41462e92649bdee7d60bd2ba83014adb" Oct 04 04:59:30 crc kubenswrapper[4770]: I1004 04:59:30.394929 4770 scope.go:117] "RemoveContainer" containerID="f9fe0391b3d07589167f36be94e5b4f1f4953d5b7029d383994b41bf2d87f1d7" Oct 04 04:59:30 crc kubenswrapper[4770]: I1004 04:59:30.888066 4770 generic.go:334] "Generic (PLEG): container finished" podID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerID="48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce" exitCode=0 Oct 04 04:59:30 crc kubenswrapper[4770]: I1004 04:59:30.888115 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wgg4m" event={"ID":"38911d7b-b75c-40bf-8ab0-296e1324e78f","Type":"ContainerDied","Data":"48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce"} Oct 04 04:59:30 crc kubenswrapper[4770]: I1004 04:59:30.891197 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tp8f7" event={"ID":"e95792e6-508b-4cb8-bc3e-35c8825c9761","Type":"ContainerStarted","Data":"dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd"} Oct 04 04:59:32 crc kubenswrapper[4770]: I1004 04:59:32.918966 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wgg4m" event={"ID":"38911d7b-b75c-40bf-8ab0-296e1324e78f","Type":"ContainerStarted","Data":"a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6"} Oct 04 04:59:37 crc kubenswrapper[4770]: I1004 04:59:37.980542 4770 generic.go:334] "Generic (PLEG): container finished" podID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerID="a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6" exitCode=0 Oct 04 04:59:37 crc kubenswrapper[4770]: I1004 04:59:37.980621 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wgg4m" event={"ID":"38911d7b-b75c-40bf-8ab0-296e1324e78f","Type":"ContainerDied","Data":"a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6"} Oct 04 04:59:40 crc kubenswrapper[4770]: I1004 04:59:40.008604 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wgg4m" event={"ID":"38911d7b-b75c-40bf-8ab0-296e1324e78f","Type":"ContainerStarted","Data":"6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde"} Oct 04 04:59:40 crc kubenswrapper[4770]: I1004 04:59:40.043702 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wgg4m" podStartSLOduration=3.793747954 podStartE2EDuration="12.043675963s" podCreationTimestamp="2025-10-04 04:59:28 +0000 UTC" firstStartedPulling="2025-10-04 04:59:30.889810241 +0000 UTC m=+6982.181819953" lastFinishedPulling="2025-10-04 04:59:39.13973825 +0000 UTC m=+6990.431747962" observedRunningTime="2025-10-04 04:59:40.035493488 +0000 UTC m=+6991.327503240" watchObservedRunningTime="2025-10-04 04:59:40.043675963 +0000 UTC m=+6991.335685695" Oct 04 04:59:41 crc kubenswrapper[4770]: I1004 04:59:41.673773 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:59:41 crc kubenswrapper[4770]: E1004 04:59:41.674503 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:59:42 crc kubenswrapper[4770]: I1004 04:59:42.027120 4770 generic.go:334] "Generic (PLEG): container finished" podID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerID="dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd" exitCode=0 Oct 04 04:59:42 crc kubenswrapper[4770]: I1004 04:59:42.027176 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tp8f7" event={"ID":"e95792e6-508b-4cb8-bc3e-35c8825c9761","Type":"ContainerDied","Data":"dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd"} Oct 04 04:59:43 crc kubenswrapper[4770]: I1004 04:59:43.042193 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tp8f7" event={"ID":"e95792e6-508b-4cb8-bc3e-35c8825c9761","Type":"ContainerStarted","Data":"edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5"} Oct 04 04:59:43 crc kubenswrapper[4770]: I1004 04:59:43.059980 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tp8f7" podStartSLOduration=3.442390163 podStartE2EDuration="17.059959084s" podCreationTimestamp="2025-10-04 04:59:26 +0000 UTC" firstStartedPulling="2025-10-04 04:59:28.842347119 +0000 UTC m=+6980.134356841" lastFinishedPulling="2025-10-04 04:59:42.45991605 +0000 UTC m=+6993.751925762" observedRunningTime="2025-10-04 04:59:43.057925041 +0000 UTC m=+6994.349934763" watchObservedRunningTime="2025-10-04 04:59:43.059959084 +0000 UTC m=+6994.351968806" Oct 04 04:59:46 crc kubenswrapper[4770]: I1004 04:59:46.732219 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:46 crc kubenswrapper[4770]: I1004 04:59:46.732844 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 04:59:47 crc kubenswrapper[4770]: I1004 04:59:47.782991 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tp8f7" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:47 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:47 crc kubenswrapper[4770]: > Oct 04 04:59:49 crc kubenswrapper[4770]: I1004 04:59:49.321856 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:49 crc kubenswrapper[4770]: I1004 04:59:49.322217 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:50 crc kubenswrapper[4770]: I1004 04:59:50.375923 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wgg4m" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:50 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:50 crc kubenswrapper[4770]: > Oct 04 04:59:52 crc kubenswrapper[4770]: I1004 04:59:52.673772 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 04:59:52 crc kubenswrapper[4770]: E1004 04:59:52.674825 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 04:59:57 crc kubenswrapper[4770]: I1004 04:59:57.779653 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tp8f7" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="registry-server" probeResult="failure" output=< Oct 04 04:59:57 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 04:59:57 crc kubenswrapper[4770]: > Oct 04 04:59:59 crc kubenswrapper[4770]: I1004 04:59:59.379102 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wgg4m" Oct 04 04:59:59 crc kubenswrapper[4770]: I1004 04:59:59.443282 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wgg4m" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.174133 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs"] Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.176416 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.179741 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.180442 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.188274 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs"] Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.206058 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wgg4m"] Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.221107 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pjxc\" (UniqueName: \"kubernetes.io/projected/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-kube-api-access-2pjxc\") pod \"collect-profiles-29325900-zjdfs\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.221203 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-secret-volume\") pod \"collect-profiles-29325900-zjdfs\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.221344 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-config-volume\") pod \"collect-profiles-29325900-zjdfs\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.322845 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pjxc\" (UniqueName: \"kubernetes.io/projected/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-kube-api-access-2pjxc\") pod \"collect-profiles-29325900-zjdfs\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.322922 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-secret-volume\") pod \"collect-profiles-29325900-zjdfs\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.323106 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-config-volume\") pod \"collect-profiles-29325900-zjdfs\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.324531 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-config-volume\") pod \"collect-profiles-29325900-zjdfs\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.330262 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-secret-volume\") pod \"collect-profiles-29325900-zjdfs\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.341311 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pjxc\" (UniqueName: \"kubernetes.io/projected/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-kube-api-access-2pjxc\") pod \"collect-profiles-29325900-zjdfs\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:00 crc kubenswrapper[4770]: I1004 05:00:00.514409 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.008445 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs"] Oct 04 05:00:01 crc kubenswrapper[4770]: W1004 05:00:01.017402 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2274a58e_0eb7_459a_8cf9_568c0f6ca8b5.slice/crio-fb8dfd06ed0362f85fb77d2c9f9f2f16d9db876ee9ff6265203d33a5dcb2ecee WatchSource:0}: Error finding container fb8dfd06ed0362f85fb77d2c9f9f2f16d9db876ee9ff6265203d33a5dcb2ecee: Status 404 returned error can't find the container with id fb8dfd06ed0362f85fb77d2c9f9f2f16d9db876ee9ff6265203d33a5dcb2ecee Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.233978 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" event={"ID":"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5","Type":"ContainerStarted","Data":"0b4cb397514a2569fc50577730957c6ded11b44120b1cea50fae539c8b096c0a"} Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.234058 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" event={"ID":"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5","Type":"ContainerStarted","Data":"fb8dfd06ed0362f85fb77d2c9f9f2f16d9db876ee9ff6265203d33a5dcb2ecee"} Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.234081 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wgg4m" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerName="registry-server" containerID="cri-o://6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde" gracePeriod=2 Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.253097 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" podStartSLOduration=1.2530752299999999 podStartE2EDuration="1.25307523s" podCreationTimestamp="2025-10-04 05:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:00:01.251663633 +0000 UTC m=+7012.543673355" watchObservedRunningTime="2025-10-04 05:00:01.25307523 +0000 UTC m=+7012.545084962" Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.791413 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wgg4m" Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.874206 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fqsk\" (UniqueName: \"kubernetes.io/projected/38911d7b-b75c-40bf-8ab0-296e1324e78f-kube-api-access-7fqsk\") pod \"38911d7b-b75c-40bf-8ab0-296e1324e78f\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.874759 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-catalog-content\") pod \"38911d7b-b75c-40bf-8ab0-296e1324e78f\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.874916 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-utilities\") pod \"38911d7b-b75c-40bf-8ab0-296e1324e78f\" (UID: \"38911d7b-b75c-40bf-8ab0-296e1324e78f\") " Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.875451 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-utilities" (OuterVolumeSpecName: "utilities") pod "38911d7b-b75c-40bf-8ab0-296e1324e78f" (UID: "38911d7b-b75c-40bf-8ab0-296e1324e78f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.875711 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.882554 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38911d7b-b75c-40bf-8ab0-296e1324e78f-kube-api-access-7fqsk" (OuterVolumeSpecName: "kube-api-access-7fqsk") pod "38911d7b-b75c-40bf-8ab0-296e1324e78f" (UID: "38911d7b-b75c-40bf-8ab0-296e1324e78f"). InnerVolumeSpecName "kube-api-access-7fqsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.927045 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38911d7b-b75c-40bf-8ab0-296e1324e78f" (UID: "38911d7b-b75c-40bf-8ab0-296e1324e78f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.977990 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38911d7b-b75c-40bf-8ab0-296e1324e78f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:01 crc kubenswrapper[4770]: I1004 05:00:01.978130 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fqsk\" (UniqueName: \"kubernetes.io/projected/38911d7b-b75c-40bf-8ab0-296e1324e78f-kube-api-access-7fqsk\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.245120 4770 generic.go:334] "Generic (PLEG): container finished" podID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerID="6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde" exitCode=0 Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.245170 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wgg4m" event={"ID":"38911d7b-b75c-40bf-8ab0-296e1324e78f","Type":"ContainerDied","Data":"6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde"} Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.245243 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wgg4m" event={"ID":"38911d7b-b75c-40bf-8ab0-296e1324e78f","Type":"ContainerDied","Data":"c424ce5c391dac115c36ed90f20f7d6c4ba072134d539f113385aa4d4558daf8"} Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.245251 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wgg4m" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.245271 4770 scope.go:117] "RemoveContainer" containerID="6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.248396 4770 generic.go:334] "Generic (PLEG): container finished" podID="2274a58e-0eb7-459a-8cf9-568c0f6ca8b5" containerID="0b4cb397514a2569fc50577730957c6ded11b44120b1cea50fae539c8b096c0a" exitCode=0 Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.248466 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" event={"ID":"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5","Type":"ContainerDied","Data":"0b4cb397514a2569fc50577730957c6ded11b44120b1cea50fae539c8b096c0a"} Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.289998 4770 scope.go:117] "RemoveContainer" containerID="a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.295381 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wgg4m"] Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.304885 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wgg4m"] Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.315211 4770 scope.go:117] "RemoveContainer" containerID="48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.363062 4770 scope.go:117] "RemoveContainer" containerID="6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde" Oct 04 05:00:02 crc kubenswrapper[4770]: E1004 05:00:02.363523 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde\": container with ID starting with 6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde not found: ID does not exist" containerID="6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.363555 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde"} err="failed to get container status \"6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde\": rpc error: code = NotFound desc = could not find container \"6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde\": container with ID starting with 6b6462b3cee69d88e4dcb21a1502a34f2bffa80a5023269f0d136e3930c58fde not found: ID does not exist" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.363577 4770 scope.go:117] "RemoveContainer" containerID="a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6" Oct 04 05:00:02 crc kubenswrapper[4770]: E1004 05:00:02.364080 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6\": container with ID starting with a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6 not found: ID does not exist" containerID="a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.364126 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6"} err="failed to get container status \"a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6\": rpc error: code = NotFound desc = could not find container \"a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6\": container with ID starting with a56c2cb7df1f2cf1d794d4cca0cbd4e2c9c538bd4f1cefc2c7a00d84a9524db6 not found: ID does not exist" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.364141 4770 scope.go:117] "RemoveContainer" containerID="48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce" Oct 04 05:00:02 crc kubenswrapper[4770]: E1004 05:00:02.365595 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce\": container with ID starting with 48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce not found: ID does not exist" containerID="48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce" Oct 04 05:00:02 crc kubenswrapper[4770]: I1004 05:00:02.365651 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce"} err="failed to get container status \"48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce\": rpc error: code = NotFound desc = could not find container \"48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce\": container with ID starting with 48a7c665b420a6a5af3f43e2ff57851044e48223166dfcfb37e8496c49cb41ce not found: ID does not exist" Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.655807 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.697246 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" path="/var/lib/kubelet/pods/38911d7b-b75c-40bf-8ab0-296e1324e78f/volumes" Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.718438 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-secret-volume\") pod \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.718486 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-config-volume\") pod \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.718560 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pjxc\" (UniqueName: \"kubernetes.io/projected/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-kube-api-access-2pjxc\") pod \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\" (UID: \"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5\") " Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.720631 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-config-volume" (OuterVolumeSpecName: "config-volume") pod "2274a58e-0eb7-459a-8cf9-568c0f6ca8b5" (UID: "2274a58e-0eb7-459a-8cf9-568c0f6ca8b5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.724571 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-kube-api-access-2pjxc" (OuterVolumeSpecName: "kube-api-access-2pjxc") pod "2274a58e-0eb7-459a-8cf9-568c0f6ca8b5" (UID: "2274a58e-0eb7-459a-8cf9-568c0f6ca8b5"). InnerVolumeSpecName "kube-api-access-2pjxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.727252 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2274a58e-0eb7-459a-8cf9-568c0f6ca8b5" (UID: "2274a58e-0eb7-459a-8cf9-568c0f6ca8b5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.820932 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.820981 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:03 crc kubenswrapper[4770]: I1004 05:00:03.820996 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pjxc\" (UniqueName: \"kubernetes.io/projected/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5-kube-api-access-2pjxc\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:04 crc kubenswrapper[4770]: I1004 05:00:04.271641 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" event={"ID":"2274a58e-0eb7-459a-8cf9-568c0f6ca8b5","Type":"ContainerDied","Data":"fb8dfd06ed0362f85fb77d2c9f9f2f16d9db876ee9ff6265203d33a5dcb2ecee"} Oct 04 05:00:04 crc kubenswrapper[4770]: I1004 05:00:04.271697 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb8dfd06ed0362f85fb77d2c9f9f2f16d9db876ee9ff6265203d33a5dcb2ecee" Oct 04 05:00:04 crc kubenswrapper[4770]: I1004 05:00:04.271771 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs" Oct 04 05:00:04 crc kubenswrapper[4770]: I1004 05:00:04.358388 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57"] Oct 04 05:00:04 crc kubenswrapper[4770]: I1004 05:00:04.368107 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325855-zvz57"] Oct 04 05:00:05 crc kubenswrapper[4770]: I1004 05:00:05.694369 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a4e6a8e-cf53-46f1-8726-f59fb2549eab" path="/var/lib/kubelet/pods/0a4e6a8e-cf53-46f1-8726-f59fb2549eab/volumes" Oct 04 05:00:06 crc kubenswrapper[4770]: I1004 05:00:06.673985 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 05:00:06 crc kubenswrapper[4770]: E1004 05:00:06.674330 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:00:07 crc kubenswrapper[4770]: I1004 05:00:07.798850 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tp8f7" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="registry-server" probeResult="failure" output=< Oct 04 05:00:07 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:00:07 crc kubenswrapper[4770]: > Oct 04 05:00:17 crc kubenswrapper[4770]: I1004 05:00:17.789459 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tp8f7" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="registry-server" probeResult="failure" output=< Oct 04 05:00:17 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:00:17 crc kubenswrapper[4770]: > Oct 04 05:00:20 crc kubenswrapper[4770]: I1004 05:00:20.675944 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 05:00:20 crc kubenswrapper[4770]: E1004 05:00:20.676670 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:00:27 crc kubenswrapper[4770]: I1004 05:00:27.790309 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tp8f7" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="registry-server" probeResult="failure" output=< Oct 04 05:00:27 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:00:27 crc kubenswrapper[4770]: > Oct 04 05:00:30 crc kubenswrapper[4770]: I1004 05:00:30.548387 4770 scope.go:117] "RemoveContainer" containerID="fbc38938c3c501a7c5ce02468bd533ba2afe4b8b4e8b2b44853e690fd57a2774" Oct 04 05:00:34 crc kubenswrapper[4770]: I1004 05:00:34.673927 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 05:00:34 crc kubenswrapper[4770]: E1004 05:00:34.675042 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:00:36 crc kubenswrapper[4770]: I1004 05:00:36.794386 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 05:00:36 crc kubenswrapper[4770]: I1004 05:00:36.849687 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 05:00:37 crc kubenswrapper[4770]: I1004 05:00:37.029112 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tp8f7"] Oct 04 05:00:38 crc kubenswrapper[4770]: I1004 05:00:38.679466 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tp8f7" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="registry-server" containerID="cri-o://edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5" gracePeriod=2 Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.188377 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.309720 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnj2q\" (UniqueName: \"kubernetes.io/projected/e95792e6-508b-4cb8-bc3e-35c8825c9761-kube-api-access-pnj2q\") pod \"e95792e6-508b-4cb8-bc3e-35c8825c9761\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.311258 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-catalog-content\") pod \"e95792e6-508b-4cb8-bc3e-35c8825c9761\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.311389 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-utilities\") pod \"e95792e6-508b-4cb8-bc3e-35c8825c9761\" (UID: \"e95792e6-508b-4cb8-bc3e-35c8825c9761\") " Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.312105 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-utilities" (OuterVolumeSpecName: "utilities") pod "e95792e6-508b-4cb8-bc3e-35c8825c9761" (UID: "e95792e6-508b-4cb8-bc3e-35c8825c9761"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.313055 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.316876 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e95792e6-508b-4cb8-bc3e-35c8825c9761-kube-api-access-pnj2q" (OuterVolumeSpecName: "kube-api-access-pnj2q") pod "e95792e6-508b-4cb8-bc3e-35c8825c9761" (UID: "e95792e6-508b-4cb8-bc3e-35c8825c9761"). InnerVolumeSpecName "kube-api-access-pnj2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.396233 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e95792e6-508b-4cb8-bc3e-35c8825c9761" (UID: "e95792e6-508b-4cb8-bc3e-35c8825c9761"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.415627 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e95792e6-508b-4cb8-bc3e-35c8825c9761-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.415682 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnj2q\" (UniqueName: \"kubernetes.io/projected/e95792e6-508b-4cb8-bc3e-35c8825c9761-kube-api-access-pnj2q\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.696335 4770 generic.go:334] "Generic (PLEG): container finished" podID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerID="edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5" exitCode=0 Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.698693 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tp8f7" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.701754 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tp8f7" event={"ID":"e95792e6-508b-4cb8-bc3e-35c8825c9761","Type":"ContainerDied","Data":"edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5"} Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.701863 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tp8f7" event={"ID":"e95792e6-508b-4cb8-bc3e-35c8825c9761","Type":"ContainerDied","Data":"edd3dc98200ed8fa0c6c9287b373c8d4167ea1c8f421a607c46a7adb09f11cee"} Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.701915 4770 scope.go:117] "RemoveContainer" containerID="edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.749546 4770 scope.go:117] "RemoveContainer" containerID="dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.750287 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tp8f7"] Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.759702 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tp8f7"] Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.790638 4770 scope.go:117] "RemoveContainer" containerID="aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.839188 4770 scope.go:117] "RemoveContainer" containerID="edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5" Oct 04 05:00:39 crc kubenswrapper[4770]: E1004 05:00:39.839687 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5\": container with ID starting with edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5 not found: ID does not exist" containerID="edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.839732 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5"} err="failed to get container status \"edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5\": rpc error: code = NotFound desc = could not find container \"edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5\": container with ID starting with edfe40cf3a7bcaac02188ffb8b02813f46548eff6cd3b1ce3a664d4264f78ce5 not found: ID does not exist" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.839768 4770 scope.go:117] "RemoveContainer" containerID="dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd" Oct 04 05:00:39 crc kubenswrapper[4770]: E1004 05:00:39.840588 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd\": container with ID starting with dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd not found: ID does not exist" containerID="dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.840612 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd"} err="failed to get container status \"dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd\": rpc error: code = NotFound desc = could not find container \"dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd\": container with ID starting with dbf0ba70b8ed899bf1a835ea120970695256e35f473fd3a179cb84f5aee056bd not found: ID does not exist" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.840653 4770 scope.go:117] "RemoveContainer" containerID="aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe" Oct 04 05:00:39 crc kubenswrapper[4770]: E1004 05:00:39.841177 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe\": container with ID starting with aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe not found: ID does not exist" containerID="aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe" Oct 04 05:00:39 crc kubenswrapper[4770]: I1004 05:00:39.841207 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe"} err="failed to get container status \"aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe\": rpc error: code = NotFound desc = could not find container \"aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe\": container with ID starting with aab11e5f7bc016211bf182d4bbf85f366b0bd9f74b6c7f3c920e37b3062a5dbe not found: ID does not exist" Oct 04 05:00:41 crc kubenswrapper[4770]: I1004 05:00:41.689078 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" path="/var/lib/kubelet/pods/e95792e6-508b-4cb8-bc3e-35c8825c9761/volumes" Oct 04 05:00:47 crc kubenswrapper[4770]: I1004 05:00:47.674387 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 05:00:47 crc kubenswrapper[4770]: E1004 05:00:47.676456 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.169206 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325901-8rxql"] Oct 04 05:01:00 crc kubenswrapper[4770]: E1004 05:01:00.170319 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerName="extract-utilities" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170341 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerName="extract-utilities" Oct 04 05:01:00 crc kubenswrapper[4770]: E1004 05:01:00.170378 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerName="registry-server" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170387 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerName="registry-server" Oct 04 05:01:00 crc kubenswrapper[4770]: E1004 05:01:00.170399 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="extract-utilities" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170408 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="extract-utilities" Oct 04 05:01:00 crc kubenswrapper[4770]: E1004 05:01:00.170426 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="registry-server" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170434 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="registry-server" Oct 04 05:01:00 crc kubenswrapper[4770]: E1004 05:01:00.170448 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerName="extract-content" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170456 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerName="extract-content" Oct 04 05:01:00 crc kubenswrapper[4770]: E1004 05:01:00.170475 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="extract-content" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170484 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="extract-content" Oct 04 05:01:00 crc kubenswrapper[4770]: E1004 05:01:00.170499 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2274a58e-0eb7-459a-8cf9-568c0f6ca8b5" containerName="collect-profiles" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170507 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="2274a58e-0eb7-459a-8cf9-568c0f6ca8b5" containerName="collect-profiles" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170781 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e95792e6-508b-4cb8-bc3e-35c8825c9761" containerName="registry-server" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170798 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="2274a58e-0eb7-459a-8cf9-568c0f6ca8b5" containerName="collect-profiles" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.170818 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="38911d7b-b75c-40bf-8ab0-296e1324e78f" containerName="registry-server" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.171814 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.180423 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325901-8rxql"] Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.288497 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-combined-ca-bundle\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.288967 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4sxs\" (UniqueName: \"kubernetes.io/projected/ff136520-ed4e-478b-9ac5-69b54366cf0e-kube-api-access-t4sxs\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.289068 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-config-data\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.289261 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-fernet-keys\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.391051 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4sxs\" (UniqueName: \"kubernetes.io/projected/ff136520-ed4e-478b-9ac5-69b54366cf0e-kube-api-access-t4sxs\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.391117 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-config-data\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.391192 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-fernet-keys\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.391246 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-combined-ca-bundle\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.397612 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-config-data\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.397900 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-fernet-keys\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.400685 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-combined-ca-bundle\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.411729 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4sxs\" (UniqueName: \"kubernetes.io/projected/ff136520-ed4e-478b-9ac5-69b54366cf0e-kube-api-access-t4sxs\") pod \"keystone-cron-29325901-8rxql\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.513135 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:00 crc kubenswrapper[4770]: I1004 05:01:00.970721 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325901-8rxql"] Oct 04 05:01:01 crc kubenswrapper[4770]: I1004 05:01:01.674259 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 05:01:01 crc kubenswrapper[4770]: E1004 05:01:01.674652 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:01:01 crc kubenswrapper[4770]: I1004 05:01:01.957457 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-8rxql" event={"ID":"ff136520-ed4e-478b-9ac5-69b54366cf0e","Type":"ContainerStarted","Data":"4e3ec9ebcb37bd8ac79b816006b351057015854e27a1ff56de5e5e1c66d208f7"} Oct 04 05:01:01 crc kubenswrapper[4770]: I1004 05:01:01.957503 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-8rxql" event={"ID":"ff136520-ed4e-478b-9ac5-69b54366cf0e","Type":"ContainerStarted","Data":"78dab1c930b020859f8adf783d0a5f6bffca7f6aaa7f31f769c6328ad5727b72"} Oct 04 05:01:01 crc kubenswrapper[4770]: I1004 05:01:01.979354 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325901-8rxql" podStartSLOduration=1.979337879 podStartE2EDuration="1.979337879s" podCreationTimestamp="2025-10-04 05:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:01:01.977487427 +0000 UTC m=+7073.269497139" watchObservedRunningTime="2025-10-04 05:01:01.979337879 +0000 UTC m=+7073.271347591" Oct 04 05:01:04 crc kubenswrapper[4770]: I1004 05:01:04.991877 4770 generic.go:334] "Generic (PLEG): container finished" podID="ff136520-ed4e-478b-9ac5-69b54366cf0e" containerID="4e3ec9ebcb37bd8ac79b816006b351057015854e27a1ff56de5e5e1c66d208f7" exitCode=0 Oct 04 05:01:04 crc kubenswrapper[4770]: I1004 05:01:04.991952 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-8rxql" event={"ID":"ff136520-ed4e-478b-9ac5-69b54366cf0e","Type":"ContainerDied","Data":"4e3ec9ebcb37bd8ac79b816006b351057015854e27a1ff56de5e5e1c66d208f7"} Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.387527 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.440606 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-fernet-keys\") pod \"ff136520-ed4e-478b-9ac5-69b54366cf0e\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.440657 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4sxs\" (UniqueName: \"kubernetes.io/projected/ff136520-ed4e-478b-9ac5-69b54366cf0e-kube-api-access-t4sxs\") pod \"ff136520-ed4e-478b-9ac5-69b54366cf0e\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.441794 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-config-data\") pod \"ff136520-ed4e-478b-9ac5-69b54366cf0e\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.441832 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-combined-ca-bundle\") pod \"ff136520-ed4e-478b-9ac5-69b54366cf0e\" (UID: \"ff136520-ed4e-478b-9ac5-69b54366cf0e\") " Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.448297 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ff136520-ed4e-478b-9ac5-69b54366cf0e" (UID: "ff136520-ed4e-478b-9ac5-69b54366cf0e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.448445 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff136520-ed4e-478b-9ac5-69b54366cf0e-kube-api-access-t4sxs" (OuterVolumeSpecName: "kube-api-access-t4sxs") pod "ff136520-ed4e-478b-9ac5-69b54366cf0e" (UID: "ff136520-ed4e-478b-9ac5-69b54366cf0e"). InnerVolumeSpecName "kube-api-access-t4sxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.480803 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff136520-ed4e-478b-9ac5-69b54366cf0e" (UID: "ff136520-ed4e-478b-9ac5-69b54366cf0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.510251 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-config-data" (OuterVolumeSpecName: "config-data") pod "ff136520-ed4e-478b-9ac5-69b54366cf0e" (UID: "ff136520-ed4e-478b-9ac5-69b54366cf0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.543580 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.543633 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4sxs\" (UniqueName: \"kubernetes.io/projected/ff136520-ed4e-478b-9ac5-69b54366cf0e-kube-api-access-t4sxs\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.543648 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:06 crc kubenswrapper[4770]: I1004 05:01:06.543658 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff136520-ed4e-478b-9ac5-69b54366cf0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:07 crc kubenswrapper[4770]: I1004 05:01:07.011456 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-8rxql" event={"ID":"ff136520-ed4e-478b-9ac5-69b54366cf0e","Type":"ContainerDied","Data":"78dab1c930b020859f8adf783d0a5f6bffca7f6aaa7f31f769c6328ad5727b72"} Oct 04 05:01:07 crc kubenswrapper[4770]: I1004 05:01:07.011853 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78dab1c930b020859f8adf783d0a5f6bffca7f6aaa7f31f769c6328ad5727b72" Oct 04 05:01:07 crc kubenswrapper[4770]: I1004 05:01:07.011532 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-8rxql" Oct 04 05:01:12 crc kubenswrapper[4770]: I1004 05:01:12.674159 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 05:01:12 crc kubenswrapper[4770]: E1004 05:01:12.675056 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:01:23 crc kubenswrapper[4770]: I1004 05:01:23.052115 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-4t2xt"] Oct 04 05:01:23 crc kubenswrapper[4770]: I1004 05:01:23.063517 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-4t2xt"] Oct 04 05:01:23 crc kubenswrapper[4770]: I1004 05:01:23.688424 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca7fd2c-4805-497e-bbdf-fecba44d2a76" path="/var/lib/kubelet/pods/1ca7fd2c-4805-497e-bbdf-fecba44d2a76/volumes" Oct 04 05:01:24 crc kubenswrapper[4770]: I1004 05:01:24.673527 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 05:01:24 crc kubenswrapper[4770]: E1004 05:01:24.674249 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:01:30 crc kubenswrapper[4770]: I1004 05:01:30.648083 4770 scope.go:117] "RemoveContainer" containerID="4af590e786535fc2185c53502157d9d8d98b8c4728abb2de52d7a300a0c7d0b2" Oct 04 05:01:34 crc kubenswrapper[4770]: I1004 05:01:34.042700 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-dcf1-account-create-zzpdm"] Oct 04 05:01:34 crc kubenswrapper[4770]: I1004 05:01:34.055650 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-dcf1-account-create-zzpdm"] Oct 04 05:01:35 crc kubenswrapper[4770]: I1004 05:01:35.674282 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 05:01:35 crc kubenswrapper[4770]: I1004 05:01:35.684714 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7759865b-1f02-4d89-b903-9e49cf7c96ee" path="/var/lib/kubelet/pods/7759865b-1f02-4d89-b903-9e49cf7c96ee/volumes" Oct 04 05:01:36 crc kubenswrapper[4770]: I1004 05:01:36.310081 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"a269f8e90e01ca64af059f50645ee2e6429da2466bd6830c4e09c203a8cea8a0"} Oct 04 05:01:50 crc kubenswrapper[4770]: I1004 05:01:50.030760 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-tmj47"] Oct 04 05:01:50 crc kubenswrapper[4770]: I1004 05:01:50.049353 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-tmj47"] Oct 04 05:01:51 crc kubenswrapper[4770]: I1004 05:01:51.708584 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3875920-d1d5-455b-b27e-49599ac8bba9" path="/var/lib/kubelet/pods/c3875920-d1d5-455b-b27e-49599ac8bba9/volumes" Oct 04 05:02:30 crc kubenswrapper[4770]: I1004 05:02:30.735073 4770 scope.go:117] "RemoveContainer" containerID="a76bd61b3e309b80e7aa0132849b648cddadc00565f44e1ebab95bdfd2837864" Oct 04 05:02:30 crc kubenswrapper[4770]: I1004 05:02:30.771807 4770 scope.go:117] "RemoveContainer" containerID="6e7bb09c62432c4846dab60223583c09739c12e22a67c5ff62f76ac504d853a2" Oct 04 05:02:36 crc kubenswrapper[4770]: I1004 05:02:36.055022 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-2xr9v"] Oct 04 05:02:36 crc kubenswrapper[4770]: I1004 05:02:36.065773 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-2xr9v"] Oct 04 05:02:37 crc kubenswrapper[4770]: I1004 05:02:37.685650 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9380ab45-8702-4380-aaae-31908686f28f" path="/var/lib/kubelet/pods/9380ab45-8702-4380-aaae-31908686f28f/volumes" Oct 04 05:02:46 crc kubenswrapper[4770]: I1004 05:02:46.053304 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-9178-account-create-4xk4p"] Oct 04 05:02:46 crc kubenswrapper[4770]: I1004 05:02:46.068491 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-9178-account-create-4xk4p"] Oct 04 05:02:47 crc kubenswrapper[4770]: I1004 05:02:47.685539 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7593e8ae-831e-4e1a-8ae0-81eea8df72b6" path="/var/lib/kubelet/pods/7593e8ae-831e-4e1a-8ae0-81eea8df72b6/volumes" Oct 04 05:03:11 crc kubenswrapper[4770]: I1004 05:03:11.058590 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-ntrkf"] Oct 04 05:03:11 crc kubenswrapper[4770]: I1004 05:03:11.068163 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-ntrkf"] Oct 04 05:03:11 crc kubenswrapper[4770]: I1004 05:03:11.688645 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b" path="/var/lib/kubelet/pods/36d7ce0d-3ed1-4b1c-bd3e-f2e0d2796f5b/volumes" Oct 04 05:03:30 crc kubenswrapper[4770]: I1004 05:03:30.943264 4770 scope.go:117] "RemoveContainer" containerID="f26c3eb368d2df57e8aae433521d2572669abe880afb4ab7eedaaed028d37ce3" Oct 04 05:03:30 crc kubenswrapper[4770]: I1004 05:03:30.968138 4770 scope.go:117] "RemoveContainer" containerID="62e15a9865e1f8a92c17193409c2b4307c78d0c95f87813c067b2711cffcc250" Oct 04 05:03:31 crc kubenswrapper[4770]: I1004 05:03:31.021262 4770 scope.go:117] "RemoveContainer" containerID="6c97583d8738f2420c30f7f149d23254031d7f8ff8d974eb300b15f247d595ad" Oct 04 05:04:01 crc kubenswrapper[4770]: I1004 05:04:01.796137 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:04:01 crc kubenswrapper[4770]: I1004 05:04:01.796701 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:04:31 crc kubenswrapper[4770]: I1004 05:04:31.795621 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:04:31 crc kubenswrapper[4770]: I1004 05:04:31.796250 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:05:01 crc kubenswrapper[4770]: I1004 05:05:01.796402 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:05:01 crc kubenswrapper[4770]: I1004 05:05:01.798401 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:05:01 crc kubenswrapper[4770]: I1004 05:05:01.798503 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:05:01 crc kubenswrapper[4770]: I1004 05:05:01.799539 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a269f8e90e01ca64af059f50645ee2e6429da2466bd6830c4e09c203a8cea8a0"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:05:01 crc kubenswrapper[4770]: I1004 05:05:01.799617 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://a269f8e90e01ca64af059f50645ee2e6429da2466bd6830c4e09c203a8cea8a0" gracePeriod=600 Oct 04 05:05:02 crc kubenswrapper[4770]: I1004 05:05:02.442654 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="a269f8e90e01ca64af059f50645ee2e6429da2466bd6830c4e09c203a8cea8a0" exitCode=0 Oct 04 05:05:02 crc kubenswrapper[4770]: I1004 05:05:02.442731 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"a269f8e90e01ca64af059f50645ee2e6429da2466bd6830c4e09c203a8cea8a0"} Oct 04 05:05:02 crc kubenswrapper[4770]: I1004 05:05:02.443381 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253"} Oct 04 05:05:02 crc kubenswrapper[4770]: I1004 05:05:02.443410 4770 scope.go:117] "RemoveContainer" containerID="a79f7feb03d60501da05a74d7175fed462717e9b83136213831b9fcd39cebaac" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.252768 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7dfkl"] Oct 04 05:05:14 crc kubenswrapper[4770]: E1004 05:05:14.253664 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff136520-ed4e-478b-9ac5-69b54366cf0e" containerName="keystone-cron" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.253677 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff136520-ed4e-478b-9ac5-69b54366cf0e" containerName="keystone-cron" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.253900 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff136520-ed4e-478b-9ac5-69b54366cf0e" containerName="keystone-cron" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.255629 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.261998 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dfkl"] Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.423816 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98sfw\" (UniqueName: \"kubernetes.io/projected/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-kube-api-access-98sfw\") pod \"certified-operators-7dfkl\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.423875 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-utilities\") pod \"certified-operators-7dfkl\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.423902 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-catalog-content\") pod \"certified-operators-7dfkl\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.526294 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98sfw\" (UniqueName: \"kubernetes.io/projected/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-kube-api-access-98sfw\") pod \"certified-operators-7dfkl\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.526683 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-utilities\") pod \"certified-operators-7dfkl\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.526720 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-catalog-content\") pod \"certified-operators-7dfkl\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.527274 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-utilities\") pod \"certified-operators-7dfkl\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.527327 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-catalog-content\") pod \"certified-operators-7dfkl\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.559101 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98sfw\" (UniqueName: \"kubernetes.io/projected/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-kube-api-access-98sfw\") pod \"certified-operators-7dfkl\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:14 crc kubenswrapper[4770]: I1004 05:05:14.621665 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:15 crc kubenswrapper[4770]: I1004 05:05:15.179177 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dfkl"] Oct 04 05:05:15 crc kubenswrapper[4770]: I1004 05:05:15.580740 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfkl" event={"ID":"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527","Type":"ContainerStarted","Data":"3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd"} Oct 04 05:05:15 crc kubenswrapper[4770]: I1004 05:05:15.581113 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfkl" event={"ID":"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527","Type":"ContainerStarted","Data":"62ad8dd26dd0dd21f4978edd6f1beb14d7883295c84191bfc2a437fe66792648"} Oct 04 05:05:16 crc kubenswrapper[4770]: I1004 05:05:16.590911 4770 generic.go:334] "Generic (PLEG): container finished" podID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerID="3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd" exitCode=0 Oct 04 05:05:16 crc kubenswrapper[4770]: I1004 05:05:16.590976 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfkl" event={"ID":"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527","Type":"ContainerDied","Data":"3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd"} Oct 04 05:05:16 crc kubenswrapper[4770]: I1004 05:05:16.600853 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:05:18 crc kubenswrapper[4770]: I1004 05:05:18.609513 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfkl" event={"ID":"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527","Type":"ContainerStarted","Data":"45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4"} Oct 04 05:05:19 crc kubenswrapper[4770]: I1004 05:05:19.622725 4770 generic.go:334] "Generic (PLEG): container finished" podID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerID="45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4" exitCode=0 Oct 04 05:05:19 crc kubenswrapper[4770]: I1004 05:05:19.622815 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfkl" event={"ID":"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527","Type":"ContainerDied","Data":"45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4"} Oct 04 05:05:22 crc kubenswrapper[4770]: I1004 05:05:22.659721 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfkl" event={"ID":"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527","Type":"ContainerStarted","Data":"910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5"} Oct 04 05:05:22 crc kubenswrapper[4770]: I1004 05:05:22.682259 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7dfkl" podStartSLOduration=3.6934357479999997 podStartE2EDuration="8.682240169s" podCreationTimestamp="2025-10-04 05:05:14 +0000 UTC" firstStartedPulling="2025-10-04 05:05:16.600571892 +0000 UTC m=+7327.892581604" lastFinishedPulling="2025-10-04 05:05:21.589376313 +0000 UTC m=+7332.881386025" observedRunningTime="2025-10-04 05:05:22.676860179 +0000 UTC m=+7333.968869891" watchObservedRunningTime="2025-10-04 05:05:22.682240169 +0000 UTC m=+7333.974249881" Oct 04 05:05:24 crc kubenswrapper[4770]: I1004 05:05:24.622195 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:24 crc kubenswrapper[4770]: I1004 05:05:24.622733 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:24 crc kubenswrapper[4770]: I1004 05:05:24.682373 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:34 crc kubenswrapper[4770]: I1004 05:05:34.686492 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:34 crc kubenswrapper[4770]: I1004 05:05:34.735233 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dfkl"] Oct 04 05:05:34 crc kubenswrapper[4770]: I1004 05:05:34.802214 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7dfkl" podUID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerName="registry-server" containerID="cri-o://910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5" gracePeriod=2 Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.276301 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.385074 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-catalog-content\") pod \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.385231 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98sfw\" (UniqueName: \"kubernetes.io/projected/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-kube-api-access-98sfw\") pod \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.385491 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-utilities\") pod \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\" (UID: \"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527\") " Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.386785 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-utilities" (OuterVolumeSpecName: "utilities") pod "ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" (UID: "ee28bc8e-32a0-4be6-a0a2-8da0f68cb527"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.391536 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-kube-api-access-98sfw" (OuterVolumeSpecName: "kube-api-access-98sfw") pod "ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" (UID: "ee28bc8e-32a0-4be6-a0a2-8da0f68cb527"). InnerVolumeSpecName "kube-api-access-98sfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.425259 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" (UID: "ee28bc8e-32a0-4be6-a0a2-8da0f68cb527"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.488709 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.488745 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.488756 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98sfw\" (UniqueName: \"kubernetes.io/projected/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527-kube-api-access-98sfw\") on node \"crc\" DevicePath \"\"" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.814056 4770 generic.go:334] "Generic (PLEG): container finished" podID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerID="910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5" exitCode=0 Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.814120 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfkl" event={"ID":"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527","Type":"ContainerDied","Data":"910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5"} Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.814477 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dfkl" event={"ID":"ee28bc8e-32a0-4be6-a0a2-8da0f68cb527","Type":"ContainerDied","Data":"62ad8dd26dd0dd21f4978edd6f1beb14d7883295c84191bfc2a437fe66792648"} Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.814507 4770 scope.go:117] "RemoveContainer" containerID="910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.814138 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dfkl" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.840944 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dfkl"] Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.841427 4770 scope.go:117] "RemoveContainer" containerID="45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.850196 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7dfkl"] Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.861196 4770 scope.go:117] "RemoveContainer" containerID="3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.907219 4770 scope.go:117] "RemoveContainer" containerID="910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5" Oct 04 05:05:35 crc kubenswrapper[4770]: E1004 05:05:35.907647 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5\": container with ID starting with 910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5 not found: ID does not exist" containerID="910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.907687 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5"} err="failed to get container status \"910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5\": rpc error: code = NotFound desc = could not find container \"910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5\": container with ID starting with 910ef9dd3f559c7e58b327b1ff82adda54622b3127f068a38103a16a732477c5 not found: ID does not exist" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.907708 4770 scope.go:117] "RemoveContainer" containerID="45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4" Oct 04 05:05:35 crc kubenswrapper[4770]: E1004 05:05:35.908934 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4\": container with ID starting with 45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4 not found: ID does not exist" containerID="45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.908975 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4"} err="failed to get container status \"45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4\": rpc error: code = NotFound desc = could not find container \"45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4\": container with ID starting with 45f9b564454616005049a268eaeb8b66f1c915083ebf6ad4508474c30e6c9aa4 not found: ID does not exist" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.909016 4770 scope.go:117] "RemoveContainer" containerID="3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd" Oct 04 05:05:35 crc kubenswrapper[4770]: E1004 05:05:35.909421 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd\": container with ID starting with 3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd not found: ID does not exist" containerID="3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd" Oct 04 05:05:35 crc kubenswrapper[4770]: I1004 05:05:35.909449 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd"} err="failed to get container status \"3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd\": rpc error: code = NotFound desc = could not find container \"3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd\": container with ID starting with 3f913d7b283c4e585aa13a7aaac24b0ea216c7f24f93858ee27e202cf2ecddbd not found: ID does not exist" Oct 04 05:05:37 crc kubenswrapper[4770]: I1004 05:05:37.684201 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" path="/var/lib/kubelet/pods/ee28bc8e-32a0-4be6-a0a2-8da0f68cb527/volumes" Oct 04 05:06:46 crc kubenswrapper[4770]: I1004 05:06:46.538423 4770 generic.go:334] "Generic (PLEG): container finished" podID="b5dcbbe1-7d4d-4c82-a705-f367cb129883" containerID="2250a68a23d625dd77b74e9e040ec40210433d1a04fafcd3c2da6f9eb2992c39" exitCode=0 Oct 04 05:06:46 crc kubenswrapper[4770]: I1004 05:06:46.538532 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" event={"ID":"b5dcbbe1-7d4d-4c82-a705-f367cb129883","Type":"ContainerDied","Data":"2250a68a23d625dd77b74e9e040ec40210433d1a04fafcd3c2da6f9eb2992c39"} Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.050346 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.232425 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-inventory\") pod \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.232528 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgbfd\" (UniqueName: \"kubernetes.io/projected/b5dcbbe1-7d4d-4c82-a705-f367cb129883-kube-api-access-rgbfd\") pod \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.232591 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ceph\") pod \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.232779 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ssh-key\") pod \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.233045 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-tripleo-cleanup-combined-ca-bundle\") pod \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\" (UID: \"b5dcbbe1-7d4d-4c82-a705-f367cb129883\") " Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.241274 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "b5dcbbe1-7d4d-4c82-a705-f367cb129883" (UID: "b5dcbbe1-7d4d-4c82-a705-f367cb129883"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.241287 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5dcbbe1-7d4d-4c82-a705-f367cb129883-kube-api-access-rgbfd" (OuterVolumeSpecName: "kube-api-access-rgbfd") pod "b5dcbbe1-7d4d-4c82-a705-f367cb129883" (UID: "b5dcbbe1-7d4d-4c82-a705-f367cb129883"). InnerVolumeSpecName "kube-api-access-rgbfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.241704 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ceph" (OuterVolumeSpecName: "ceph") pod "b5dcbbe1-7d4d-4c82-a705-f367cb129883" (UID: "b5dcbbe1-7d4d-4c82-a705-f367cb129883"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.267779 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-inventory" (OuterVolumeSpecName: "inventory") pod "b5dcbbe1-7d4d-4c82-a705-f367cb129883" (UID: "b5dcbbe1-7d4d-4c82-a705-f367cb129883"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.273874 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b5dcbbe1-7d4d-4c82-a705-f367cb129883" (UID: "b5dcbbe1-7d4d-4c82-a705-f367cb129883"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.335468 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgbfd\" (UniqueName: \"kubernetes.io/projected/b5dcbbe1-7d4d-4c82-a705-f367cb129883-kube-api-access-rgbfd\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.335502 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.335512 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.335521 4770 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.335531 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5dcbbe1-7d4d-4c82-a705-f367cb129883-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.558595 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" event={"ID":"b5dcbbe1-7d4d-4c82-a705-f367cb129883","Type":"ContainerDied","Data":"bd5ef5e5325343ea89d4c4a571656be2a461cc1f710efa4b03a27f02b8bdd1e9"} Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.558642 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd5ef5e5325343ea89d4c4a571656be2a461cc1f710efa4b03a27f02b8bdd1e9" Oct 04 05:06:48 crc kubenswrapper[4770]: I1004 05:06:48.558733 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.582378 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-7gd5m"] Oct 04 05:06:55 crc kubenswrapper[4770]: E1004 05:06:55.583797 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5dcbbe1-7d4d-4c82-a705-f367cb129883" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.583826 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5dcbbe1-7d4d-4c82-a705-f367cb129883" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 04 05:06:55 crc kubenswrapper[4770]: E1004 05:06:55.583868 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerName="extract-content" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.583886 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerName="extract-content" Oct 04 05:06:55 crc kubenswrapper[4770]: E1004 05:06:55.583921 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerName="registry-server" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.583936 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerName="registry-server" Oct 04 05:06:55 crc kubenswrapper[4770]: E1004 05:06:55.583996 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerName="extract-utilities" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.584041 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerName="extract-utilities" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.584492 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5dcbbe1-7d4d-4c82-a705-f367cb129883" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.584522 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee28bc8e-32a0-4be6-a0a2-8da0f68cb527" containerName="registry-server" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.586287 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.588733 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.590145 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.590724 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.590817 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.598888 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-7gd5m"] Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.694579 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.694631 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhstl\" (UniqueName: \"kubernetes.io/projected/55f43d2c-35b3-491a-a080-dee559ff5e84-kube-api-access-vhstl\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.694659 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.694689 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-inventory\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.694756 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ceph\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.798277 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.798359 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhstl\" (UniqueName: \"kubernetes.io/projected/55f43d2c-35b3-491a-a080-dee559ff5e84-kube-api-access-vhstl\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.798898 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.799077 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-inventory\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.799404 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ceph\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.804607 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ceph\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.806164 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.806315 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-inventory\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.809417 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.822179 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhstl\" (UniqueName: \"kubernetes.io/projected/55f43d2c-35b3-491a-a080-dee559ff5e84-kube-api-access-vhstl\") pod \"bootstrap-openstack-openstack-cell1-7gd5m\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:55 crc kubenswrapper[4770]: I1004 05:06:55.911458 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:06:56 crc kubenswrapper[4770]: I1004 05:06:56.537320 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-7gd5m"] Oct 04 05:06:56 crc kubenswrapper[4770]: I1004 05:06:56.638519 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" event={"ID":"55f43d2c-35b3-491a-a080-dee559ff5e84","Type":"ContainerStarted","Data":"fe12193de486074ac55a1684d8d55e29c42de2de4eb643a51b72531f262922f3"} Oct 04 05:06:57 crc kubenswrapper[4770]: I1004 05:06:57.651775 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" event={"ID":"55f43d2c-35b3-491a-a080-dee559ff5e84","Type":"ContainerStarted","Data":"1112e6f977d889d9d3e8b09787ef40fc0322d2cf667341339cb9581650471480"} Oct 04 05:06:57 crc kubenswrapper[4770]: I1004 05:06:57.680212 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" podStartSLOduration=2.236235495 podStartE2EDuration="2.680194373s" podCreationTimestamp="2025-10-04 05:06:55 +0000 UTC" firstStartedPulling="2025-10-04 05:06:56.543346194 +0000 UTC m=+7427.835355906" lastFinishedPulling="2025-10-04 05:06:56.987305072 +0000 UTC m=+7428.279314784" observedRunningTime="2025-10-04 05:06:57.673675882 +0000 UTC m=+7428.965685624" watchObservedRunningTime="2025-10-04 05:06:57.680194373 +0000 UTC m=+7428.972204085" Oct 04 05:07:31 crc kubenswrapper[4770]: I1004 05:07:31.795391 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:07:31 crc kubenswrapper[4770]: I1004 05:07:31.795926 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:08:01 crc kubenswrapper[4770]: I1004 05:08:01.796641 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:08:01 crc kubenswrapper[4770]: I1004 05:08:01.797904 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:08:31 crc kubenswrapper[4770]: I1004 05:08:31.795891 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:08:31 crc kubenswrapper[4770]: I1004 05:08:31.796468 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:08:31 crc kubenswrapper[4770]: I1004 05:08:31.796522 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:08:31 crc kubenswrapper[4770]: I1004 05:08:31.797505 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:08:31 crc kubenswrapper[4770]: I1004 05:08:31.797583 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" gracePeriod=600 Oct 04 05:08:31 crc kubenswrapper[4770]: E1004 05:08:31.958105 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:08:32 crc kubenswrapper[4770]: I1004 05:08:32.702326 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" exitCode=0 Oct 04 05:08:32 crc kubenswrapper[4770]: I1004 05:08:32.702391 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253"} Oct 04 05:08:32 crc kubenswrapper[4770]: I1004 05:08:32.702439 4770 scope.go:117] "RemoveContainer" containerID="a269f8e90e01ca64af059f50645ee2e6429da2466bd6830c4e09c203a8cea8a0" Oct 04 05:08:32 crc kubenswrapper[4770]: I1004 05:08:32.703315 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:08:32 crc kubenswrapper[4770]: E1004 05:08:32.703762 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.195520 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-58gqm"] Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.198107 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.216911 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-58gqm"] Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.305456 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-utilities\") pod \"redhat-marketplace-58gqm\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.305609 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-catalog-content\") pod \"redhat-marketplace-58gqm\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.305676 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjnhx\" (UniqueName: \"kubernetes.io/projected/af666fa6-c0d7-4d90-af67-ec7ad604d28c-kube-api-access-zjnhx\") pod \"redhat-marketplace-58gqm\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.407795 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-utilities\") pod \"redhat-marketplace-58gqm\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.408227 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-catalog-content\") pod \"redhat-marketplace-58gqm\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.408283 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjnhx\" (UniqueName: \"kubernetes.io/projected/af666fa6-c0d7-4d90-af67-ec7ad604d28c-kube-api-access-zjnhx\") pod \"redhat-marketplace-58gqm\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.408309 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-utilities\") pod \"redhat-marketplace-58gqm\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.408676 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-catalog-content\") pod \"redhat-marketplace-58gqm\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.434209 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjnhx\" (UniqueName: \"kubernetes.io/projected/af666fa6-c0d7-4d90-af67-ec7ad604d28c-kube-api-access-zjnhx\") pod \"redhat-marketplace-58gqm\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.519288 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:43 crc kubenswrapper[4770]: I1004 05:08:43.675297 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:08:43 crc kubenswrapper[4770]: E1004 05:08:43.676141 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:08:44 crc kubenswrapper[4770]: I1004 05:08:44.010501 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-58gqm"] Oct 04 05:08:44 crc kubenswrapper[4770]: I1004 05:08:44.838958 4770 generic.go:334] "Generic (PLEG): container finished" podID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerID="6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f" exitCode=0 Oct 04 05:08:44 crc kubenswrapper[4770]: I1004 05:08:44.839063 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58gqm" event={"ID":"af666fa6-c0d7-4d90-af67-ec7ad604d28c","Type":"ContainerDied","Data":"6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f"} Oct 04 05:08:44 crc kubenswrapper[4770]: I1004 05:08:44.839300 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58gqm" event={"ID":"af666fa6-c0d7-4d90-af67-ec7ad604d28c","Type":"ContainerStarted","Data":"4ee44a7406b77aa7a4116374964ea830ba8850c7c0f7e542ab4a99d6ab206fcb"} Oct 04 05:08:45 crc kubenswrapper[4770]: I1004 05:08:45.851424 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58gqm" event={"ID":"af666fa6-c0d7-4d90-af67-ec7ad604d28c","Type":"ContainerStarted","Data":"186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59"} Oct 04 05:08:46 crc kubenswrapper[4770]: I1004 05:08:46.865001 4770 generic.go:334] "Generic (PLEG): container finished" podID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerID="186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59" exitCode=0 Oct 04 05:08:46 crc kubenswrapper[4770]: I1004 05:08:46.865215 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58gqm" event={"ID":"af666fa6-c0d7-4d90-af67-ec7ad604d28c","Type":"ContainerDied","Data":"186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59"} Oct 04 05:08:48 crc kubenswrapper[4770]: I1004 05:08:48.895502 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58gqm" event={"ID":"af666fa6-c0d7-4d90-af67-ec7ad604d28c","Type":"ContainerStarted","Data":"9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8"} Oct 04 05:08:48 crc kubenswrapper[4770]: I1004 05:08:48.924256 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-58gqm" podStartSLOduration=2.99233147 podStartE2EDuration="5.924234414s" podCreationTimestamp="2025-10-04 05:08:43 +0000 UTC" firstStartedPulling="2025-10-04 05:08:44.841461001 +0000 UTC m=+7536.133470713" lastFinishedPulling="2025-10-04 05:08:47.773363935 +0000 UTC m=+7539.065373657" observedRunningTime="2025-10-04 05:08:48.916826678 +0000 UTC m=+7540.208836440" watchObservedRunningTime="2025-10-04 05:08:48.924234414 +0000 UTC m=+7540.216244136" Oct 04 05:08:53 crc kubenswrapper[4770]: I1004 05:08:53.520500 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:53 crc kubenswrapper[4770]: I1004 05:08:53.521224 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:53 crc kubenswrapper[4770]: I1004 05:08:53.592551 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:54 crc kubenswrapper[4770]: I1004 05:08:54.012479 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:54 crc kubenswrapper[4770]: I1004 05:08:54.071141 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-58gqm"] Oct 04 05:08:54 crc kubenswrapper[4770]: I1004 05:08:54.674316 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:08:54 crc kubenswrapper[4770]: E1004 05:08:54.674691 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:08:55 crc kubenswrapper[4770]: I1004 05:08:55.979967 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-58gqm" podUID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerName="registry-server" containerID="cri-o://9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8" gracePeriod=2 Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.588562 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.715211 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-catalog-content\") pod \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.715412 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjnhx\" (UniqueName: \"kubernetes.io/projected/af666fa6-c0d7-4d90-af67-ec7ad604d28c-kube-api-access-zjnhx\") pod \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.715448 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-utilities\") pod \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\" (UID: \"af666fa6-c0d7-4d90-af67-ec7ad604d28c\") " Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.716506 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-utilities" (OuterVolumeSpecName: "utilities") pod "af666fa6-c0d7-4d90-af67-ec7ad604d28c" (UID: "af666fa6-c0d7-4d90-af67-ec7ad604d28c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.726362 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af666fa6-c0d7-4d90-af67-ec7ad604d28c-kube-api-access-zjnhx" (OuterVolumeSpecName: "kube-api-access-zjnhx") pod "af666fa6-c0d7-4d90-af67-ec7ad604d28c" (UID: "af666fa6-c0d7-4d90-af67-ec7ad604d28c"). InnerVolumeSpecName "kube-api-access-zjnhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.735393 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af666fa6-c0d7-4d90-af67-ec7ad604d28c" (UID: "af666fa6-c0d7-4d90-af67-ec7ad604d28c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.818779 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.818835 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjnhx\" (UniqueName: \"kubernetes.io/projected/af666fa6-c0d7-4d90-af67-ec7ad604d28c-kube-api-access-zjnhx\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.818850 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af666fa6-c0d7-4d90-af67-ec7ad604d28c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.995816 4770 generic.go:334] "Generic (PLEG): container finished" podID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerID="9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8" exitCode=0 Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.995857 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58gqm" Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.995880 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58gqm" event={"ID":"af666fa6-c0d7-4d90-af67-ec7ad604d28c","Type":"ContainerDied","Data":"9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8"} Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.995918 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58gqm" event={"ID":"af666fa6-c0d7-4d90-af67-ec7ad604d28c","Type":"ContainerDied","Data":"4ee44a7406b77aa7a4116374964ea830ba8850c7c0f7e542ab4a99d6ab206fcb"} Oct 04 05:08:56 crc kubenswrapper[4770]: I1004 05:08:56.995938 4770 scope.go:117] "RemoveContainer" containerID="9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8" Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.022660 4770 scope.go:117] "RemoveContainer" containerID="186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59" Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.034673 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-58gqm"] Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.043255 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-58gqm"] Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.064187 4770 scope.go:117] "RemoveContainer" containerID="6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f" Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.103842 4770 scope.go:117] "RemoveContainer" containerID="9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8" Oct 04 05:08:57 crc kubenswrapper[4770]: E1004 05:08:57.104708 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8\": container with ID starting with 9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8 not found: ID does not exist" containerID="9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8" Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.104810 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8"} err="failed to get container status \"9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8\": rpc error: code = NotFound desc = could not find container \"9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8\": container with ID starting with 9bc2d00ae518402144b3aa33f95e143ba3ce8435ca1aba0acdd3d4aef25237b8 not found: ID does not exist" Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.104856 4770 scope.go:117] "RemoveContainer" containerID="186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59" Oct 04 05:08:57 crc kubenswrapper[4770]: E1004 05:08:57.105482 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59\": container with ID starting with 186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59 not found: ID does not exist" containerID="186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59" Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.105548 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59"} err="failed to get container status \"186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59\": rpc error: code = NotFound desc = could not find container \"186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59\": container with ID starting with 186b133bcbb8a9af7bfa2c8516c13007fc6f237a6a10eba3fdab5ee4f5153b59 not found: ID does not exist" Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.105595 4770 scope.go:117] "RemoveContainer" containerID="6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f" Oct 04 05:08:57 crc kubenswrapper[4770]: E1004 05:08:57.106158 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f\": container with ID starting with 6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f not found: ID does not exist" containerID="6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f" Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.106202 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f"} err="failed to get container status \"6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f\": rpc error: code = NotFound desc = could not find container \"6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f\": container with ID starting with 6bc6893b6201921e367e38094b01db174b52bb7ac98bbc8dc126d684ec239a8f not found: ID does not exist" Oct 04 05:08:57 crc kubenswrapper[4770]: I1004 05:08:57.709225 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" path="/var/lib/kubelet/pods/af666fa6-c0d7-4d90-af67-ec7ad604d28c/volumes" Oct 04 05:09:08 crc kubenswrapper[4770]: I1004 05:09:08.678310 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:09:08 crc kubenswrapper[4770]: E1004 05:09:08.679572 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:09:19 crc kubenswrapper[4770]: I1004 05:09:19.680390 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:09:19 crc kubenswrapper[4770]: E1004 05:09:19.681326 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:09:31 crc kubenswrapper[4770]: I1004 05:09:31.676041 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:09:31 crc kubenswrapper[4770]: E1004 05:09:31.677095 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.138262 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9zzcm"] Oct 04 05:09:39 crc kubenswrapper[4770]: E1004 05:09:39.140342 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerName="registry-server" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.140440 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerName="registry-server" Oct 04 05:09:39 crc kubenswrapper[4770]: E1004 05:09:39.140534 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerName="extract-content" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.140605 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerName="extract-content" Oct 04 05:09:39 crc kubenswrapper[4770]: E1004 05:09:39.140684 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerName="extract-utilities" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.140746 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerName="extract-utilities" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.141000 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="af666fa6-c0d7-4d90-af67-ec7ad604d28c" containerName="registry-server" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.143117 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.159653 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zzcm"] Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.234720 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6wnl\" (UniqueName: \"kubernetes.io/projected/99466da4-81c6-4460-9c6d-080d3457edcc-kube-api-access-d6wnl\") pod \"community-operators-9zzcm\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.235232 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-utilities\") pod \"community-operators-9zzcm\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.235331 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-catalog-content\") pod \"community-operators-9zzcm\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.337512 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-utilities\") pod \"community-operators-9zzcm\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.337764 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-catalog-content\") pod \"community-operators-9zzcm\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.337932 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6wnl\" (UniqueName: \"kubernetes.io/projected/99466da4-81c6-4460-9c6d-080d3457edcc-kube-api-access-d6wnl\") pod \"community-operators-9zzcm\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.337948 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-utilities\") pod \"community-operators-9zzcm\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.338160 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-catalog-content\") pod \"community-operators-9zzcm\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.362866 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6wnl\" (UniqueName: \"kubernetes.io/projected/99466da4-81c6-4460-9c6d-080d3457edcc-kube-api-access-d6wnl\") pod \"community-operators-9zzcm\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:39 crc kubenswrapper[4770]: I1004 05:09:39.478786 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:40 crc kubenswrapper[4770]: I1004 05:09:40.162827 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zzcm"] Oct 04 05:09:40 crc kubenswrapper[4770]: I1004 05:09:40.554195 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zzcm" event={"ID":"99466da4-81c6-4460-9c6d-080d3457edcc","Type":"ContainerStarted","Data":"7bc991c2ef6077dafab86e0be2d1bfbc43aece0619bd6aba7c76104949203229"} Oct 04 05:09:40 crc kubenswrapper[4770]: I1004 05:09:40.554552 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zzcm" event={"ID":"99466da4-81c6-4460-9c6d-080d3457edcc","Type":"ContainerStarted","Data":"1940b2ceafafbd4c395479d7345f13cc41feae5cc107eae93b24557b3b1180af"} Oct 04 05:09:41 crc kubenswrapper[4770]: I1004 05:09:41.569228 4770 generic.go:334] "Generic (PLEG): container finished" podID="99466da4-81c6-4460-9c6d-080d3457edcc" containerID="7bc991c2ef6077dafab86e0be2d1bfbc43aece0619bd6aba7c76104949203229" exitCode=0 Oct 04 05:09:41 crc kubenswrapper[4770]: I1004 05:09:41.569303 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zzcm" event={"ID":"99466da4-81c6-4460-9c6d-080d3457edcc","Type":"ContainerDied","Data":"7bc991c2ef6077dafab86e0be2d1bfbc43aece0619bd6aba7c76104949203229"} Oct 04 05:09:42 crc kubenswrapper[4770]: I1004 05:09:42.580529 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zzcm" event={"ID":"99466da4-81c6-4460-9c6d-080d3457edcc","Type":"ContainerStarted","Data":"010e0458886d5ea64f04f5f881872cd10d6fb70d23fc564fbf462c91b5b28bb6"} Oct 04 05:09:42 crc kubenswrapper[4770]: I1004 05:09:42.673816 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:09:42 crc kubenswrapper[4770]: E1004 05:09:42.674155 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:09:44 crc kubenswrapper[4770]: I1004 05:09:44.600642 4770 generic.go:334] "Generic (PLEG): container finished" podID="99466da4-81c6-4460-9c6d-080d3457edcc" containerID="010e0458886d5ea64f04f5f881872cd10d6fb70d23fc564fbf462c91b5b28bb6" exitCode=0 Oct 04 05:09:44 crc kubenswrapper[4770]: I1004 05:09:44.600679 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zzcm" event={"ID":"99466da4-81c6-4460-9c6d-080d3457edcc","Type":"ContainerDied","Data":"010e0458886d5ea64f04f5f881872cd10d6fb70d23fc564fbf462c91b5b28bb6"} Oct 04 05:09:46 crc kubenswrapper[4770]: I1004 05:09:46.623958 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zzcm" event={"ID":"99466da4-81c6-4460-9c6d-080d3457edcc","Type":"ContainerStarted","Data":"dea2502c4d14ad7b8cf534d79544d6e3c6cf8c9762318ff3bb918c1fdd69ee0a"} Oct 04 05:09:46 crc kubenswrapper[4770]: I1004 05:09:46.652537 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9zzcm" podStartSLOduration=3.793473954 podStartE2EDuration="7.652515793s" podCreationTimestamp="2025-10-04 05:09:39 +0000 UTC" firstStartedPulling="2025-10-04 05:09:41.572598543 +0000 UTC m=+7592.864608295" lastFinishedPulling="2025-10-04 05:09:45.431640422 +0000 UTC m=+7596.723650134" observedRunningTime="2025-10-04 05:09:46.645470829 +0000 UTC m=+7597.937480541" watchObservedRunningTime="2025-10-04 05:09:46.652515793 +0000 UTC m=+7597.944525515" Oct 04 05:09:49 crc kubenswrapper[4770]: I1004 05:09:49.479368 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:49 crc kubenswrapper[4770]: I1004 05:09:49.479709 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:49 crc kubenswrapper[4770]: I1004 05:09:49.530274 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:56 crc kubenswrapper[4770]: I1004 05:09:56.674661 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:09:56 crc kubenswrapper[4770]: E1004 05:09:56.675945 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:09:59 crc kubenswrapper[4770]: I1004 05:09:59.557234 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:09:59 crc kubenswrapper[4770]: I1004 05:09:59.621824 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zzcm"] Oct 04 05:09:59 crc kubenswrapper[4770]: I1004 05:09:59.774632 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9zzcm" podUID="99466da4-81c6-4460-9c6d-080d3457edcc" containerName="registry-server" containerID="cri-o://dea2502c4d14ad7b8cf534d79544d6e3c6cf8c9762318ff3bb918c1fdd69ee0a" gracePeriod=2 Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.786759 4770 generic.go:334] "Generic (PLEG): container finished" podID="99466da4-81c6-4460-9c6d-080d3457edcc" containerID="dea2502c4d14ad7b8cf534d79544d6e3c6cf8c9762318ff3bb918c1fdd69ee0a" exitCode=0 Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.786810 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zzcm" event={"ID":"99466da4-81c6-4460-9c6d-080d3457edcc","Type":"ContainerDied","Data":"dea2502c4d14ad7b8cf534d79544d6e3c6cf8c9762318ff3bb918c1fdd69ee0a"} Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.787114 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zzcm" event={"ID":"99466da4-81c6-4460-9c6d-080d3457edcc","Type":"ContainerDied","Data":"1940b2ceafafbd4c395479d7345f13cc41feae5cc107eae93b24557b3b1180af"} Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.787129 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1940b2ceafafbd4c395479d7345f13cc41feae5cc107eae93b24557b3b1180af" Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.847056 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.952088 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-catalog-content\") pod \"99466da4-81c6-4460-9c6d-080d3457edcc\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.952252 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6wnl\" (UniqueName: \"kubernetes.io/projected/99466da4-81c6-4460-9c6d-080d3457edcc-kube-api-access-d6wnl\") pod \"99466da4-81c6-4460-9c6d-080d3457edcc\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.952282 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-utilities\") pod \"99466da4-81c6-4460-9c6d-080d3457edcc\" (UID: \"99466da4-81c6-4460-9c6d-080d3457edcc\") " Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.953391 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-utilities" (OuterVolumeSpecName: "utilities") pod "99466da4-81c6-4460-9c6d-080d3457edcc" (UID: "99466da4-81c6-4460-9c6d-080d3457edcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.955222 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:00 crc kubenswrapper[4770]: I1004 05:10:00.957951 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99466da4-81c6-4460-9c6d-080d3457edcc-kube-api-access-d6wnl" (OuterVolumeSpecName: "kube-api-access-d6wnl") pod "99466da4-81c6-4460-9c6d-080d3457edcc" (UID: "99466da4-81c6-4460-9c6d-080d3457edcc"). InnerVolumeSpecName "kube-api-access-d6wnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:10:01 crc kubenswrapper[4770]: I1004 05:10:01.034266 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99466da4-81c6-4460-9c6d-080d3457edcc" (UID: "99466da4-81c6-4460-9c6d-080d3457edcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:01 crc kubenswrapper[4770]: I1004 05:10:01.056890 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6wnl\" (UniqueName: \"kubernetes.io/projected/99466da4-81c6-4460-9c6d-080d3457edcc-kube-api-access-d6wnl\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:01 crc kubenswrapper[4770]: I1004 05:10:01.056936 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99466da4-81c6-4460-9c6d-080d3457edcc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:01 crc kubenswrapper[4770]: I1004 05:10:01.797187 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zzcm" Oct 04 05:10:01 crc kubenswrapper[4770]: I1004 05:10:01.822674 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zzcm"] Oct 04 05:10:01 crc kubenswrapper[4770]: I1004 05:10:01.830136 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9zzcm"] Oct 04 05:10:03 crc kubenswrapper[4770]: I1004 05:10:03.692299 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99466da4-81c6-4460-9c6d-080d3457edcc" path="/var/lib/kubelet/pods/99466da4-81c6-4460-9c6d-080d3457edcc/volumes" Oct 04 05:10:03 crc kubenswrapper[4770]: I1004 05:10:03.827690 4770 generic.go:334] "Generic (PLEG): container finished" podID="55f43d2c-35b3-491a-a080-dee559ff5e84" containerID="1112e6f977d889d9d3e8b09787ef40fc0322d2cf667341339cb9581650471480" exitCode=0 Oct 04 05:10:03 crc kubenswrapper[4770]: I1004 05:10:03.827760 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" event={"ID":"55f43d2c-35b3-491a-a080-dee559ff5e84","Type":"ContainerDied","Data":"1112e6f977d889d9d3e8b09787ef40fc0322d2cf667341339cb9581650471480"} Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.337329 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.458318 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhstl\" (UniqueName: \"kubernetes.io/projected/55f43d2c-35b3-491a-a080-dee559ff5e84-kube-api-access-vhstl\") pod \"55f43d2c-35b3-491a-a080-dee559ff5e84\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.458362 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ceph\") pod \"55f43d2c-35b3-491a-a080-dee559ff5e84\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.458391 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-inventory\") pod \"55f43d2c-35b3-491a-a080-dee559ff5e84\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.458412 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ssh-key\") pod \"55f43d2c-35b3-491a-a080-dee559ff5e84\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.459326 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-bootstrap-combined-ca-bundle\") pod \"55f43d2c-35b3-491a-a080-dee559ff5e84\" (UID: \"55f43d2c-35b3-491a-a080-dee559ff5e84\") " Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.464493 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ceph" (OuterVolumeSpecName: "ceph") pod "55f43d2c-35b3-491a-a080-dee559ff5e84" (UID: "55f43d2c-35b3-491a-a080-dee559ff5e84"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.465123 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55f43d2c-35b3-491a-a080-dee559ff5e84-kube-api-access-vhstl" (OuterVolumeSpecName: "kube-api-access-vhstl") pod "55f43d2c-35b3-491a-a080-dee559ff5e84" (UID: "55f43d2c-35b3-491a-a080-dee559ff5e84"). InnerVolumeSpecName "kube-api-access-vhstl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.465770 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "55f43d2c-35b3-491a-a080-dee559ff5e84" (UID: "55f43d2c-35b3-491a-a080-dee559ff5e84"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.491418 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "55f43d2c-35b3-491a-a080-dee559ff5e84" (UID: "55f43d2c-35b3-491a-a080-dee559ff5e84"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.503319 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-inventory" (OuterVolumeSpecName: "inventory") pod "55f43d2c-35b3-491a-a080-dee559ff5e84" (UID: "55f43d2c-35b3-491a-a080-dee559ff5e84"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.561346 4770 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.561382 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhstl\" (UniqueName: \"kubernetes.io/projected/55f43d2c-35b3-491a-a080-dee559ff5e84-kube-api-access-vhstl\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.561391 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.561401 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.561409 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55f43d2c-35b3-491a-a080-dee559ff5e84-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.850113 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" event={"ID":"55f43d2c-35b3-491a-a080-dee559ff5e84","Type":"ContainerDied","Data":"fe12193de486074ac55a1684d8d55e29c42de2de4eb643a51b72531f262922f3"} Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.850158 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-7gd5m" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.850181 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe12193de486074ac55a1684d8d55e29c42de2de4eb643a51b72531f262922f3" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.961427 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-kxtjn"] Oct 04 05:10:05 crc kubenswrapper[4770]: E1004 05:10:05.962160 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99466da4-81c6-4460-9c6d-080d3457edcc" containerName="extract-utilities" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.962302 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="99466da4-81c6-4460-9c6d-080d3457edcc" containerName="extract-utilities" Oct 04 05:10:05 crc kubenswrapper[4770]: E1004 05:10:05.962407 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99466da4-81c6-4460-9c6d-080d3457edcc" containerName="extract-content" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.962485 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="99466da4-81c6-4460-9c6d-080d3457edcc" containerName="extract-content" Oct 04 05:10:05 crc kubenswrapper[4770]: E1004 05:10:05.962574 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99466da4-81c6-4460-9c6d-080d3457edcc" containerName="registry-server" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.962652 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="99466da4-81c6-4460-9c6d-080d3457edcc" containerName="registry-server" Oct 04 05:10:05 crc kubenswrapper[4770]: E1004 05:10:05.962750 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55f43d2c-35b3-491a-a080-dee559ff5e84" containerName="bootstrap-openstack-openstack-cell1" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.962826 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="55f43d2c-35b3-491a-a080-dee559ff5e84" containerName="bootstrap-openstack-openstack-cell1" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.963176 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="55f43d2c-35b3-491a-a080-dee559ff5e84" containerName="bootstrap-openstack-openstack-cell1" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.963293 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="99466da4-81c6-4460-9c6d-080d3457edcc" containerName="registry-server" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.964265 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.966645 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.966893 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.968521 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.969459 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:10:05 crc kubenswrapper[4770]: I1004 05:10:05.971152 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-kxtjn"] Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.072114 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-inventory\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.072215 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ssh-key\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.072267 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ceph\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.072406 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsddf\" (UniqueName: \"kubernetes.io/projected/ef4ea9df-722f-4700-b51d-17693400e22a-kube-api-access-fsddf\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.174164 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsddf\" (UniqueName: \"kubernetes.io/projected/ef4ea9df-722f-4700-b51d-17693400e22a-kube-api-access-fsddf\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.174254 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-inventory\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.174335 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ssh-key\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.174918 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ceph\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.178491 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-inventory\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.178714 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ssh-key\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.178810 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ceph\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.194627 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsddf\" (UniqueName: \"kubernetes.io/projected/ef4ea9df-722f-4700-b51d-17693400e22a-kube-api-access-fsddf\") pod \"download-cache-openstack-openstack-cell1-kxtjn\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.283054 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.816722 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-kxtjn"] Oct 04 05:10:06 crc kubenswrapper[4770]: I1004 05:10:06.874197 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" event={"ID":"ef4ea9df-722f-4700-b51d-17693400e22a","Type":"ContainerStarted","Data":"6dd26371181009cf8b7606e094f73fdaedfbe9a52c2dfbca78b9c9b249a9a566"} Oct 04 05:10:07 crc kubenswrapper[4770]: I1004 05:10:07.884656 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" event={"ID":"ef4ea9df-722f-4700-b51d-17693400e22a","Type":"ContainerStarted","Data":"befb8a2bc4c773c07b5dafecec584a10fda9cf323a593d46d2a47f37343831ef"} Oct 04 05:10:07 crc kubenswrapper[4770]: I1004 05:10:07.908149 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" podStartSLOduration=2.308624056 podStartE2EDuration="2.908131755s" podCreationTimestamp="2025-10-04 05:10:05 +0000 UTC" firstStartedPulling="2025-10-04 05:10:06.821046014 +0000 UTC m=+7618.113055726" lastFinishedPulling="2025-10-04 05:10:07.420553683 +0000 UTC m=+7618.712563425" observedRunningTime="2025-10-04 05:10:07.90027482 +0000 UTC m=+7619.192284532" watchObservedRunningTime="2025-10-04 05:10:07.908131755 +0000 UTC m=+7619.200141467" Oct 04 05:10:08 crc kubenswrapper[4770]: I1004 05:10:08.674377 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:10:08 crc kubenswrapper[4770]: E1004 05:10:08.674710 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:10:22 crc kubenswrapper[4770]: I1004 05:10:22.673862 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:10:22 crc kubenswrapper[4770]: E1004 05:10:22.674613 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:10:36 crc kubenswrapper[4770]: I1004 05:10:36.673326 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:10:36 crc kubenswrapper[4770]: E1004 05:10:36.674215 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:10:39 crc kubenswrapper[4770]: I1004 05:10:39.945704 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4xn49"] Oct 04 05:10:39 crc kubenswrapper[4770]: I1004 05:10:39.950461 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:39 crc kubenswrapper[4770]: I1004 05:10:39.954323 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xn49"] Oct 04 05:10:39 crc kubenswrapper[4770]: I1004 05:10:39.984903 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-catalog-content\") pod \"redhat-operators-4xn49\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:39 crc kubenswrapper[4770]: I1004 05:10:39.985267 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-utilities\") pod \"redhat-operators-4xn49\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:39 crc kubenswrapper[4770]: I1004 05:10:39.985614 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9594\" (UniqueName: \"kubernetes.io/projected/25faf479-fcd4-46c8-a2fd-420fc1c26c20-kube-api-access-z9594\") pod \"redhat-operators-4xn49\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:40 crc kubenswrapper[4770]: I1004 05:10:40.087663 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-utilities\") pod \"redhat-operators-4xn49\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:40 crc kubenswrapper[4770]: I1004 05:10:40.087821 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9594\" (UniqueName: \"kubernetes.io/projected/25faf479-fcd4-46c8-a2fd-420fc1c26c20-kube-api-access-z9594\") pod \"redhat-operators-4xn49\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:40 crc kubenswrapper[4770]: I1004 05:10:40.087947 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-catalog-content\") pod \"redhat-operators-4xn49\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:40 crc kubenswrapper[4770]: I1004 05:10:40.088204 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-utilities\") pod \"redhat-operators-4xn49\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:40 crc kubenswrapper[4770]: I1004 05:10:40.088469 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-catalog-content\") pod \"redhat-operators-4xn49\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:40 crc kubenswrapper[4770]: I1004 05:10:40.105409 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9594\" (UniqueName: \"kubernetes.io/projected/25faf479-fcd4-46c8-a2fd-420fc1c26c20-kube-api-access-z9594\") pod \"redhat-operators-4xn49\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:40 crc kubenswrapper[4770]: I1004 05:10:40.274044 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:10:49 crc kubenswrapper[4770]: I1004 05:10:49.653641 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="c7631708-8285-44c1-b416-ad9e0dae57b8" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Oct 04 05:10:49 crc kubenswrapper[4770]: I1004 05:10:49.819089 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:10:49 crc kubenswrapper[4770]: E1004 05:10:49.821093 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:10:50 crc kubenswrapper[4770]: I1004 05:10:50.300667 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xn49"] Oct 04 05:10:50 crc kubenswrapper[4770]: I1004 05:10:50.844462 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xn49" event={"ID":"25faf479-fcd4-46c8-a2fd-420fc1c26c20","Type":"ContainerStarted","Data":"67f6d52d12fea472e4481dba919cecea884244db3c0809247208efaec4ac7fad"} Oct 04 05:10:51 crc kubenswrapper[4770]: I1004 05:10:51.859925 4770 generic.go:334] "Generic (PLEG): container finished" podID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerID="66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535" exitCode=0 Oct 04 05:10:51 crc kubenswrapper[4770]: I1004 05:10:51.859981 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xn49" event={"ID":"25faf479-fcd4-46c8-a2fd-420fc1c26c20","Type":"ContainerDied","Data":"66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535"} Oct 04 05:10:51 crc kubenswrapper[4770]: I1004 05:10:51.864533 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:10:53 crc kubenswrapper[4770]: I1004 05:10:53.888052 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xn49" event={"ID":"25faf479-fcd4-46c8-a2fd-420fc1c26c20","Type":"ContainerStarted","Data":"df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e"} Oct 04 05:11:00 crc kubenswrapper[4770]: I1004 05:11:00.675682 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:11:00 crc kubenswrapper[4770]: E1004 05:11:00.677437 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:11:14 crc kubenswrapper[4770]: I1004 05:11:14.675848 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:11:14 crc kubenswrapper[4770]: E1004 05:11:14.677183 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:11:18 crc kubenswrapper[4770]: I1004 05:11:18.163672 4770 generic.go:334] "Generic (PLEG): container finished" podID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerID="df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e" exitCode=0 Oct 04 05:11:18 crc kubenswrapper[4770]: I1004 05:11:18.163869 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xn49" event={"ID":"25faf479-fcd4-46c8-a2fd-420fc1c26c20","Type":"ContainerDied","Data":"df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e"} Oct 04 05:11:25 crc kubenswrapper[4770]: I1004 05:11:25.673515 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:11:25 crc kubenswrapper[4770]: E1004 05:11:25.674310 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:11:35 crc kubenswrapper[4770]: I1004 05:11:35.329399 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xn49" event={"ID":"25faf479-fcd4-46c8-a2fd-420fc1c26c20","Type":"ContainerStarted","Data":"23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce"} Oct 04 05:11:35 crc kubenswrapper[4770]: I1004 05:11:35.351124 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4xn49" podStartSLOduration=13.984304988 podStartE2EDuration="56.35110602s" podCreationTimestamp="2025-10-04 05:10:39 +0000 UTC" firstStartedPulling="2025-10-04 05:10:51.864046909 +0000 UTC m=+7663.156056621" lastFinishedPulling="2025-10-04 05:11:34.230847941 +0000 UTC m=+7705.522857653" observedRunningTime="2025-10-04 05:11:35.347801274 +0000 UTC m=+7706.639810996" watchObservedRunningTime="2025-10-04 05:11:35.35110602 +0000 UTC m=+7706.643115722" Oct 04 05:11:37 crc kubenswrapper[4770]: I1004 05:11:37.676249 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:11:37 crc kubenswrapper[4770]: E1004 05:11:37.677092 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:11:40 crc kubenswrapper[4770]: I1004 05:11:40.274512 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:11:40 crc kubenswrapper[4770]: I1004 05:11:40.275802 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:11:41 crc kubenswrapper[4770]: I1004 05:11:41.333312 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4xn49" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerName="registry-server" probeResult="failure" output=< Oct 04 05:11:41 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:11:41 crc kubenswrapper[4770]: > Oct 04 05:11:49 crc kubenswrapper[4770]: I1004 05:11:49.685393 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:11:49 crc kubenswrapper[4770]: E1004 05:11:49.686255 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:11:50 crc kubenswrapper[4770]: I1004 05:11:50.344159 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:11:50 crc kubenswrapper[4770]: I1004 05:11:50.408091 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:11:50 crc kubenswrapper[4770]: I1004 05:11:50.581416 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xn49"] Oct 04 05:11:51 crc kubenswrapper[4770]: I1004 05:11:51.500517 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4xn49" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerName="registry-server" containerID="cri-o://23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce" gracePeriod=2 Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.051735 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.085260 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9594\" (UniqueName: \"kubernetes.io/projected/25faf479-fcd4-46c8-a2fd-420fc1c26c20-kube-api-access-z9594\") pod \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.085360 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-catalog-content\") pod \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.093286 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25faf479-fcd4-46c8-a2fd-420fc1c26c20-kube-api-access-z9594" (OuterVolumeSpecName: "kube-api-access-z9594") pod "25faf479-fcd4-46c8-a2fd-420fc1c26c20" (UID: "25faf479-fcd4-46c8-a2fd-420fc1c26c20"). InnerVolumeSpecName "kube-api-access-z9594". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.188667 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25faf479-fcd4-46c8-a2fd-420fc1c26c20" (UID: "25faf479-fcd4-46c8-a2fd-420fc1c26c20"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.189529 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-utilities\") pod \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\" (UID: \"25faf479-fcd4-46c8-a2fd-420fc1c26c20\") " Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.190182 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-utilities" (OuterVolumeSpecName: "utilities") pod "25faf479-fcd4-46c8-a2fd-420fc1c26c20" (UID: "25faf479-fcd4-46c8-a2fd-420fc1c26c20"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.190578 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.190599 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9594\" (UniqueName: \"kubernetes.io/projected/25faf479-fcd4-46c8-a2fd-420fc1c26c20-kube-api-access-z9594\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.190611 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25faf479-fcd4-46c8-a2fd-420fc1c26c20-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.514035 4770 generic.go:334] "Generic (PLEG): container finished" podID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerID="23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce" exitCode=0 Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.514062 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xn49" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.514060 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xn49" event={"ID":"25faf479-fcd4-46c8-a2fd-420fc1c26c20","Type":"ContainerDied","Data":"23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce"} Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.514645 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xn49" event={"ID":"25faf479-fcd4-46c8-a2fd-420fc1c26c20","Type":"ContainerDied","Data":"67f6d52d12fea472e4481dba919cecea884244db3c0809247208efaec4ac7fad"} Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.514693 4770 scope.go:117] "RemoveContainer" containerID="23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.538728 4770 scope.go:117] "RemoveContainer" containerID="df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.556299 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xn49"] Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.566929 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4xn49"] Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.590148 4770 scope.go:117] "RemoveContainer" containerID="66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.621847 4770 scope.go:117] "RemoveContainer" containerID="23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce" Oct 04 05:11:52 crc kubenswrapper[4770]: E1004 05:11:52.622462 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce\": container with ID starting with 23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce not found: ID does not exist" containerID="23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.622574 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce"} err="failed to get container status \"23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce\": rpc error: code = NotFound desc = could not find container \"23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce\": container with ID starting with 23f86839f6cd4e1bae8375a461b3a208f3492ef78da0fe07755e8b662a9311ce not found: ID does not exist" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.622742 4770 scope.go:117] "RemoveContainer" containerID="df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e" Oct 04 05:11:52 crc kubenswrapper[4770]: E1004 05:11:52.623119 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e\": container with ID starting with df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e not found: ID does not exist" containerID="df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.623145 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e"} err="failed to get container status \"df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e\": rpc error: code = NotFound desc = could not find container \"df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e\": container with ID starting with df58bb645c7de73c12fca4e373a840f91a41c47e30de7768959bfb7286470e3e not found: ID does not exist" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.623159 4770 scope.go:117] "RemoveContainer" containerID="66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535" Oct 04 05:11:52 crc kubenswrapper[4770]: E1004 05:11:52.623567 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535\": container with ID starting with 66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535 not found: ID does not exist" containerID="66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535" Oct 04 05:11:52 crc kubenswrapper[4770]: I1004 05:11:52.623669 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535"} err="failed to get container status \"66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535\": rpc error: code = NotFound desc = could not find container \"66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535\": container with ID starting with 66d7a628f9a6551e007d843cf8da049c700b27abbf8851809864735fdb835535 not found: ID does not exist" Oct 04 05:11:52 crc kubenswrapper[4770]: E1004 05:11:52.798523 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25faf479_fcd4_46c8_a2fd_420fc1c26c20.slice/crio-67f6d52d12fea472e4481dba919cecea884244db3c0809247208efaec4ac7fad\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25faf479_fcd4_46c8_a2fd_420fc1c26c20.slice\": RecentStats: unable to find data in memory cache]" Oct 04 05:11:53 crc kubenswrapper[4770]: I1004 05:11:53.690525 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" path="/var/lib/kubelet/pods/25faf479-fcd4-46c8-a2fd-420fc1c26c20/volumes" Oct 04 05:11:54 crc kubenswrapper[4770]: I1004 05:11:54.535887 4770 generic.go:334] "Generic (PLEG): container finished" podID="ef4ea9df-722f-4700-b51d-17693400e22a" containerID="befb8a2bc4c773c07b5dafecec584a10fda9cf323a593d46d2a47f37343831ef" exitCode=0 Oct 04 05:11:54 crc kubenswrapper[4770]: I1004 05:11:54.535990 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" event={"ID":"ef4ea9df-722f-4700-b51d-17693400e22a","Type":"ContainerDied","Data":"befb8a2bc4c773c07b5dafecec584a10fda9cf323a593d46d2a47f37343831ef"} Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.007092 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.208031 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-inventory\") pod \"ef4ea9df-722f-4700-b51d-17693400e22a\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.208584 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ssh-key\") pod \"ef4ea9df-722f-4700-b51d-17693400e22a\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.208619 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsddf\" (UniqueName: \"kubernetes.io/projected/ef4ea9df-722f-4700-b51d-17693400e22a-kube-api-access-fsddf\") pod \"ef4ea9df-722f-4700-b51d-17693400e22a\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.208661 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ceph\") pod \"ef4ea9df-722f-4700-b51d-17693400e22a\" (UID: \"ef4ea9df-722f-4700-b51d-17693400e22a\") " Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.213920 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ceph" (OuterVolumeSpecName: "ceph") pod "ef4ea9df-722f-4700-b51d-17693400e22a" (UID: "ef4ea9df-722f-4700-b51d-17693400e22a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.214167 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef4ea9df-722f-4700-b51d-17693400e22a-kube-api-access-fsddf" (OuterVolumeSpecName: "kube-api-access-fsddf") pod "ef4ea9df-722f-4700-b51d-17693400e22a" (UID: "ef4ea9df-722f-4700-b51d-17693400e22a"). InnerVolumeSpecName "kube-api-access-fsddf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.239912 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-inventory" (OuterVolumeSpecName: "inventory") pod "ef4ea9df-722f-4700-b51d-17693400e22a" (UID: "ef4ea9df-722f-4700-b51d-17693400e22a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.243222 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef4ea9df-722f-4700-b51d-17693400e22a" (UID: "ef4ea9df-722f-4700-b51d-17693400e22a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.311658 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.311702 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsddf\" (UniqueName: \"kubernetes.io/projected/ef4ea9df-722f-4700-b51d-17693400e22a-kube-api-access-fsddf\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.311717 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.311729 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef4ea9df-722f-4700-b51d-17693400e22a-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.556175 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" event={"ID":"ef4ea9df-722f-4700-b51d-17693400e22a","Type":"ContainerDied","Data":"6dd26371181009cf8b7606e094f73fdaedfbe9a52c2dfbca78b9c9b249a9a566"} Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.556261 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dd26371181009cf8b7606e094f73fdaedfbe9a52c2dfbca78b9c9b249a9a566" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.556343 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-kxtjn" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.647813 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-49sv8"] Oct 04 05:11:56 crc kubenswrapper[4770]: E1004 05:11:56.648446 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerName="extract-content" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.648463 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerName="extract-content" Oct 04 05:11:56 crc kubenswrapper[4770]: E1004 05:11:56.648485 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef4ea9df-722f-4700-b51d-17693400e22a" containerName="download-cache-openstack-openstack-cell1" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.648493 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef4ea9df-722f-4700-b51d-17693400e22a" containerName="download-cache-openstack-openstack-cell1" Oct 04 05:11:56 crc kubenswrapper[4770]: E1004 05:11:56.648506 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerName="extract-utilities" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.648514 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerName="extract-utilities" Oct 04 05:11:56 crc kubenswrapper[4770]: E1004 05:11:56.648538 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerName="registry-server" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.648568 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerName="registry-server" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.648818 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef4ea9df-722f-4700-b51d-17693400e22a" containerName="download-cache-openstack-openstack-cell1" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.648838 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="25faf479-fcd4-46c8-a2fd-420fc1c26c20" containerName="registry-server" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.649800 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.654542 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.654769 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.654815 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.655199 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.655950 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-49sv8"] Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.825156 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg2rf\" (UniqueName: \"kubernetes.io/projected/e23f70c8-5f06-4908-9907-f8af47aef701-kube-api-access-bg2rf\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.825776 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ceph\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.825913 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-inventory\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.826102 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ssh-key\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.929277 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg2rf\" (UniqueName: \"kubernetes.io/projected/e23f70c8-5f06-4908-9907-f8af47aef701-kube-api-access-bg2rf\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.929636 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ceph\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.929777 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-inventory\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.929914 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ssh-key\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.937901 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-inventory\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.938702 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ssh-key\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.948121 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ceph\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.955210 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg2rf\" (UniqueName: \"kubernetes.io/projected/e23f70c8-5f06-4908-9907-f8af47aef701-kube-api-access-bg2rf\") pod \"configure-network-openstack-openstack-cell1-49sv8\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:56 crc kubenswrapper[4770]: I1004 05:11:56.984680 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:11:57 crc kubenswrapper[4770]: I1004 05:11:57.511163 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-49sv8"] Oct 04 05:11:57 crc kubenswrapper[4770]: I1004 05:11:57.566195 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-49sv8" event={"ID":"e23f70c8-5f06-4908-9907-f8af47aef701","Type":"ContainerStarted","Data":"db12371cc3ae8cd1a9eac16baae128e384b74bdcded5d4828eb22fa3ff879625"} Oct 04 05:11:58 crc kubenswrapper[4770]: I1004 05:11:58.577890 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-49sv8" event={"ID":"e23f70c8-5f06-4908-9907-f8af47aef701","Type":"ContainerStarted","Data":"868d8a86c386a6aedc4c416e297f5bdccec292fcb3ca757a2d3105a2dd0f39de"} Oct 04 05:11:58 crc kubenswrapper[4770]: I1004 05:11:58.613705 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-49sv8" podStartSLOduration=2.119714382 podStartE2EDuration="2.613677091s" podCreationTimestamp="2025-10-04 05:11:56 +0000 UTC" firstStartedPulling="2025-10-04 05:11:57.515206121 +0000 UTC m=+7728.807215823" lastFinishedPulling="2025-10-04 05:11:58.00916882 +0000 UTC m=+7729.301178532" observedRunningTime="2025-10-04 05:11:58.601150013 +0000 UTC m=+7729.893159755" watchObservedRunningTime="2025-10-04 05:11:58.613677091 +0000 UTC m=+7729.905686843" Oct 04 05:12:03 crc kubenswrapper[4770]: I1004 05:12:03.674376 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:12:03 crc kubenswrapper[4770]: E1004 05:12:03.675235 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:12:16 crc kubenswrapper[4770]: I1004 05:12:16.673655 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:12:16 crc kubenswrapper[4770]: E1004 05:12:16.676248 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:12:30 crc kubenswrapper[4770]: I1004 05:12:30.674619 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:12:30 crc kubenswrapper[4770]: E1004 05:12:30.675510 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:12:45 crc kubenswrapper[4770]: I1004 05:12:45.673848 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:12:45 crc kubenswrapper[4770]: E1004 05:12:45.674688 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:12:56 crc kubenswrapper[4770]: I1004 05:12:56.674099 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:12:56 crc kubenswrapper[4770]: E1004 05:12:56.675211 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:13:10 crc kubenswrapper[4770]: I1004 05:13:10.674066 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:13:10 crc kubenswrapper[4770]: E1004 05:13:10.674928 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:13:18 crc kubenswrapper[4770]: I1004 05:13:18.458269 4770 generic.go:334] "Generic (PLEG): container finished" podID="e23f70c8-5f06-4908-9907-f8af47aef701" containerID="868d8a86c386a6aedc4c416e297f5bdccec292fcb3ca757a2d3105a2dd0f39de" exitCode=0 Oct 04 05:13:18 crc kubenswrapper[4770]: I1004 05:13:18.458405 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-49sv8" event={"ID":"e23f70c8-5f06-4908-9907-f8af47aef701","Type":"ContainerDied","Data":"868d8a86c386a6aedc4c416e297f5bdccec292fcb3ca757a2d3105a2dd0f39de"} Oct 04 05:13:19 crc kubenswrapper[4770]: I1004 05:13:19.919569 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:13:19 crc kubenswrapper[4770]: I1004 05:13:19.966643 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg2rf\" (UniqueName: \"kubernetes.io/projected/e23f70c8-5f06-4908-9907-f8af47aef701-kube-api-access-bg2rf\") pod \"e23f70c8-5f06-4908-9907-f8af47aef701\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " Oct 04 05:13:19 crc kubenswrapper[4770]: I1004 05:13:19.966833 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ssh-key\") pod \"e23f70c8-5f06-4908-9907-f8af47aef701\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " Oct 04 05:13:19 crc kubenswrapper[4770]: I1004 05:13:19.966876 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ceph\") pod \"e23f70c8-5f06-4908-9907-f8af47aef701\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " Oct 04 05:13:19 crc kubenswrapper[4770]: I1004 05:13:19.966965 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-inventory\") pod \"e23f70c8-5f06-4908-9907-f8af47aef701\" (UID: \"e23f70c8-5f06-4908-9907-f8af47aef701\") " Oct 04 05:13:19 crc kubenswrapper[4770]: I1004 05:13:19.975201 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ceph" (OuterVolumeSpecName: "ceph") pod "e23f70c8-5f06-4908-9907-f8af47aef701" (UID: "e23f70c8-5f06-4908-9907-f8af47aef701"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:19 crc kubenswrapper[4770]: I1004 05:13:19.975269 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e23f70c8-5f06-4908-9907-f8af47aef701-kube-api-access-bg2rf" (OuterVolumeSpecName: "kube-api-access-bg2rf") pod "e23f70c8-5f06-4908-9907-f8af47aef701" (UID: "e23f70c8-5f06-4908-9907-f8af47aef701"). InnerVolumeSpecName "kube-api-access-bg2rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:13:19 crc kubenswrapper[4770]: I1004 05:13:19.999522 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e23f70c8-5f06-4908-9907-f8af47aef701" (UID: "e23f70c8-5f06-4908-9907-f8af47aef701"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:19.999979 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-inventory" (OuterVolumeSpecName: "inventory") pod "e23f70c8-5f06-4908-9907-f8af47aef701" (UID: "e23f70c8-5f06-4908-9907-f8af47aef701"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.069533 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.069582 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg2rf\" (UniqueName: \"kubernetes.io/projected/e23f70c8-5f06-4908-9907-f8af47aef701-kube-api-access-bg2rf\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.069596 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.069607 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e23f70c8-5f06-4908-9907-f8af47aef701-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.502953 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-49sv8" event={"ID":"e23f70c8-5f06-4908-9907-f8af47aef701","Type":"ContainerDied","Data":"db12371cc3ae8cd1a9eac16baae128e384b74bdcded5d4828eb22fa3ff879625"} Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.503045 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db12371cc3ae8cd1a9eac16baae128e384b74bdcded5d4828eb22fa3ff879625" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.503152 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-49sv8" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.573957 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-64fpm"] Oct 04 05:13:20 crc kubenswrapper[4770]: E1004 05:13:20.574487 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23f70c8-5f06-4908-9907-f8af47aef701" containerName="configure-network-openstack-openstack-cell1" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.574505 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23f70c8-5f06-4908-9907-f8af47aef701" containerName="configure-network-openstack-openstack-cell1" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.574750 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23f70c8-5f06-4908-9907-f8af47aef701" containerName="configure-network-openstack-openstack-cell1" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.575651 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.579758 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.579952 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.580115 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.580256 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.584112 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-64fpm"] Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.686769 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ceph\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.686939 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mm28\" (UniqueName: \"kubernetes.io/projected/17d78a1f-a6ab-4523-a033-f9388abb2ada-kube-api-access-5mm28\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.687055 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-inventory\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.687105 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ssh-key\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.789445 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ceph\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.789797 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mm28\" (UniqueName: \"kubernetes.io/projected/17d78a1f-a6ab-4523-a033-f9388abb2ada-kube-api-access-5mm28\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.789868 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-inventory\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.789903 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ssh-key\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.793248 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ssh-key\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.795149 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ceph\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.797117 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-inventory\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.810550 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mm28\" (UniqueName: \"kubernetes.io/projected/17d78a1f-a6ab-4523-a033-f9388abb2ada-kube-api-access-5mm28\") pod \"validate-network-openstack-openstack-cell1-64fpm\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:20 crc kubenswrapper[4770]: I1004 05:13:20.942724 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:21 crc kubenswrapper[4770]: I1004 05:13:21.518663 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-64fpm"] Oct 04 05:13:22 crc kubenswrapper[4770]: I1004 05:13:22.532606 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-64fpm" event={"ID":"17d78a1f-a6ab-4523-a033-f9388abb2ada","Type":"ContainerStarted","Data":"05ed32a1c122c117060430a23b05037ef8da2f338383e95d990d88402731912e"} Oct 04 05:13:22 crc kubenswrapper[4770]: I1004 05:13:22.532969 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-64fpm" event={"ID":"17d78a1f-a6ab-4523-a033-f9388abb2ada","Type":"ContainerStarted","Data":"6bdf45f36492e60cd479f92fdb9f4faf21593ce9debe32bb118fa95fc60fe68b"} Oct 04 05:13:22 crc kubenswrapper[4770]: I1004 05:13:22.553995 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-64fpm" podStartSLOduration=2.123924361 podStartE2EDuration="2.553971048s" podCreationTimestamp="2025-10-04 05:13:20 +0000 UTC" firstStartedPulling="2025-10-04 05:13:21.524271137 +0000 UTC m=+7812.816280859" lastFinishedPulling="2025-10-04 05:13:21.954317834 +0000 UTC m=+7813.246327546" observedRunningTime="2025-10-04 05:13:22.549914382 +0000 UTC m=+7813.841924104" watchObservedRunningTime="2025-10-04 05:13:22.553971048 +0000 UTC m=+7813.845980800" Oct 04 05:13:23 crc kubenswrapper[4770]: I1004 05:13:23.674154 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:13:23 crc kubenswrapper[4770]: E1004 05:13:23.674671 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:13:27 crc kubenswrapper[4770]: I1004 05:13:27.609202 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-64fpm" event={"ID":"17d78a1f-a6ab-4523-a033-f9388abb2ada","Type":"ContainerDied","Data":"05ed32a1c122c117060430a23b05037ef8da2f338383e95d990d88402731912e"} Oct 04 05:13:27 crc kubenswrapper[4770]: I1004 05:13:27.609132 4770 generic.go:334] "Generic (PLEG): container finished" podID="17d78a1f-a6ab-4523-a033-f9388abb2ada" containerID="05ed32a1c122c117060430a23b05037ef8da2f338383e95d990d88402731912e" exitCode=0 Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.058522 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.084651 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mm28\" (UniqueName: \"kubernetes.io/projected/17d78a1f-a6ab-4523-a033-f9388abb2ada-kube-api-access-5mm28\") pod \"17d78a1f-a6ab-4523-a033-f9388abb2ada\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.084733 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ceph\") pod \"17d78a1f-a6ab-4523-a033-f9388abb2ada\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.084845 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-inventory\") pod \"17d78a1f-a6ab-4523-a033-f9388abb2ada\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.084928 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ssh-key\") pod \"17d78a1f-a6ab-4523-a033-f9388abb2ada\" (UID: \"17d78a1f-a6ab-4523-a033-f9388abb2ada\") " Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.093844 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ceph" (OuterVolumeSpecName: "ceph") pod "17d78a1f-a6ab-4523-a033-f9388abb2ada" (UID: "17d78a1f-a6ab-4523-a033-f9388abb2ada"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.094251 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d78a1f-a6ab-4523-a033-f9388abb2ada-kube-api-access-5mm28" (OuterVolumeSpecName: "kube-api-access-5mm28") pod "17d78a1f-a6ab-4523-a033-f9388abb2ada" (UID: "17d78a1f-a6ab-4523-a033-f9388abb2ada"). InnerVolumeSpecName "kube-api-access-5mm28". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.136171 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "17d78a1f-a6ab-4523-a033-f9388abb2ada" (UID: "17d78a1f-a6ab-4523-a033-f9388abb2ada"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.139157 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-inventory" (OuterVolumeSpecName: "inventory") pod "17d78a1f-a6ab-4523-a033-f9388abb2ada" (UID: "17d78a1f-a6ab-4523-a033-f9388abb2ada"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.186024 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.186056 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.186066 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mm28\" (UniqueName: \"kubernetes.io/projected/17d78a1f-a6ab-4523-a033-f9388abb2ada-kube-api-access-5mm28\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.186075 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d78a1f-a6ab-4523-a033-f9388abb2ada-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.214963 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-zdb69"] Oct 04 05:13:29 crc kubenswrapper[4770]: E1004 05:13:29.215715 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d78a1f-a6ab-4523-a033-f9388abb2ada" containerName="validate-network-openstack-openstack-cell1" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.215736 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d78a1f-a6ab-4523-a033-f9388abb2ada" containerName="validate-network-openstack-openstack-cell1" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.216049 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d78a1f-a6ab-4523-a033-f9388abb2ada" containerName="validate-network-openstack-openstack-cell1" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.216981 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.228192 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-zdb69"] Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.287098 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-inventory\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.287213 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ssh-key\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.287437 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ceph\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.287812 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c6v7\" (UniqueName: \"kubernetes.io/projected/8b6ca520-b8fe-491e-8a3d-307a5489c908-kube-api-access-2c6v7\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.389731 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c6v7\" (UniqueName: \"kubernetes.io/projected/8b6ca520-b8fe-491e-8a3d-307a5489c908-kube-api-access-2c6v7\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.389872 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-inventory\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.389956 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ssh-key\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.389994 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ceph\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.393661 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ceph\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.393977 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ssh-key\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.394085 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-inventory\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.411395 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c6v7\" (UniqueName: \"kubernetes.io/projected/8b6ca520-b8fe-491e-8a3d-307a5489c908-kube-api-access-2c6v7\") pod \"install-os-openstack-openstack-cell1-zdb69\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.557411 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.633028 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-64fpm" event={"ID":"17d78a1f-a6ab-4523-a033-f9388abb2ada","Type":"ContainerDied","Data":"6bdf45f36492e60cd479f92fdb9f4faf21593ce9debe32bb118fa95fc60fe68b"} Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.633477 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bdf45f36492e60cd479f92fdb9f4faf21593ce9debe32bb118fa95fc60fe68b" Oct 04 05:13:29 crc kubenswrapper[4770]: I1004 05:13:29.633188 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-64fpm" Oct 04 05:13:30 crc kubenswrapper[4770]: I1004 05:13:30.118663 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-zdb69"] Oct 04 05:13:30 crc kubenswrapper[4770]: W1004 05:13:30.123100 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b6ca520_b8fe_491e_8a3d_307a5489c908.slice/crio-9f30a0b8e644c956e6137fcf36091b9cc67f10f22251eea6680b2041a43bfa0e WatchSource:0}: Error finding container 9f30a0b8e644c956e6137fcf36091b9cc67f10f22251eea6680b2041a43bfa0e: Status 404 returned error can't find the container with id 9f30a0b8e644c956e6137fcf36091b9cc67f10f22251eea6680b2041a43bfa0e Oct 04 05:13:30 crc kubenswrapper[4770]: I1004 05:13:30.643907 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-zdb69" event={"ID":"8b6ca520-b8fe-491e-8a3d-307a5489c908","Type":"ContainerStarted","Data":"9f30a0b8e644c956e6137fcf36091b9cc67f10f22251eea6680b2041a43bfa0e"} Oct 04 05:13:31 crc kubenswrapper[4770]: I1004 05:13:31.655121 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-zdb69" event={"ID":"8b6ca520-b8fe-491e-8a3d-307a5489c908","Type":"ContainerStarted","Data":"0ca04c3d68dd19c6ab6b39c4085c2ad86a90562d4aa4e91684dae961cbf0d16d"} Oct 04 05:13:31 crc kubenswrapper[4770]: I1004 05:13:31.673884 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-zdb69" podStartSLOduration=1.698343186 podStartE2EDuration="2.673863909s" podCreationTimestamp="2025-10-04 05:13:29 +0000 UTC" firstStartedPulling="2025-10-04 05:13:30.125826873 +0000 UTC m=+7821.417836585" lastFinishedPulling="2025-10-04 05:13:31.101347576 +0000 UTC m=+7822.393357308" observedRunningTime="2025-10-04 05:13:31.669958338 +0000 UTC m=+7822.961968050" watchObservedRunningTime="2025-10-04 05:13:31.673863909 +0000 UTC m=+7822.965873621" Oct 04 05:13:38 crc kubenswrapper[4770]: I1004 05:13:38.674148 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:13:39 crc kubenswrapper[4770]: I1004 05:13:39.761877 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"4899d0932e29047d30bb5ffc6e04e928969b5942a1d3a16b4990fd1f4b9ce26f"} Oct 04 05:14:15 crc kubenswrapper[4770]: I1004 05:14:15.124164 4770 generic.go:334] "Generic (PLEG): container finished" podID="8b6ca520-b8fe-491e-8a3d-307a5489c908" containerID="0ca04c3d68dd19c6ab6b39c4085c2ad86a90562d4aa4e91684dae961cbf0d16d" exitCode=0 Oct 04 05:14:15 crc kubenswrapper[4770]: I1004 05:14:15.124247 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-zdb69" event={"ID":"8b6ca520-b8fe-491e-8a3d-307a5489c908","Type":"ContainerDied","Data":"0ca04c3d68dd19c6ab6b39c4085c2ad86a90562d4aa4e91684dae961cbf0d16d"} Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.576364 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.678932 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ssh-key\") pod \"8b6ca520-b8fe-491e-8a3d-307a5489c908\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.678996 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-inventory\") pod \"8b6ca520-b8fe-491e-8a3d-307a5489c908\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.679078 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c6v7\" (UniqueName: \"kubernetes.io/projected/8b6ca520-b8fe-491e-8a3d-307a5489c908-kube-api-access-2c6v7\") pod \"8b6ca520-b8fe-491e-8a3d-307a5489c908\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.679177 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ceph\") pod \"8b6ca520-b8fe-491e-8a3d-307a5489c908\" (UID: \"8b6ca520-b8fe-491e-8a3d-307a5489c908\") " Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.686335 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ceph" (OuterVolumeSpecName: "ceph") pod "8b6ca520-b8fe-491e-8a3d-307a5489c908" (UID: "8b6ca520-b8fe-491e-8a3d-307a5489c908"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.686518 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b6ca520-b8fe-491e-8a3d-307a5489c908-kube-api-access-2c6v7" (OuterVolumeSpecName: "kube-api-access-2c6v7") pod "8b6ca520-b8fe-491e-8a3d-307a5489c908" (UID: "8b6ca520-b8fe-491e-8a3d-307a5489c908"). InnerVolumeSpecName "kube-api-access-2c6v7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.721889 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-inventory" (OuterVolumeSpecName: "inventory") pod "8b6ca520-b8fe-491e-8a3d-307a5489c908" (UID: "8b6ca520-b8fe-491e-8a3d-307a5489c908"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.723314 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8b6ca520-b8fe-491e-8a3d-307a5489c908" (UID: "8b6ca520-b8fe-491e-8a3d-307a5489c908"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.782346 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.782388 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.782401 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c6v7\" (UniqueName: \"kubernetes.io/projected/8b6ca520-b8fe-491e-8a3d-307a5489c908-kube-api-access-2c6v7\") on node \"crc\" DevicePath \"\"" Oct 04 05:14:16 crc kubenswrapper[4770]: I1004 05:14:16.782412 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8b6ca520-b8fe-491e-8a3d-307a5489c908-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.147862 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-zdb69" event={"ID":"8b6ca520-b8fe-491e-8a3d-307a5489c908","Type":"ContainerDied","Data":"9f30a0b8e644c956e6137fcf36091b9cc67f10f22251eea6680b2041a43bfa0e"} Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.148260 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f30a0b8e644c956e6137fcf36091b9cc67f10f22251eea6680b2041a43bfa0e" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.147911 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-zdb69" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.226330 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-dv8tn"] Oct 04 05:14:17 crc kubenswrapper[4770]: E1004 05:14:17.227294 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b6ca520-b8fe-491e-8a3d-307a5489c908" containerName="install-os-openstack-openstack-cell1" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.227327 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b6ca520-b8fe-491e-8a3d-307a5489c908" containerName="install-os-openstack-openstack-cell1" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.227683 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b6ca520-b8fe-491e-8a3d-307a5489c908" containerName="install-os-openstack-openstack-cell1" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.229178 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.231570 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.231798 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.231885 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.232321 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.252569 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-dv8tn"] Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.291518 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ssh-key\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.291793 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-inventory\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.291876 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glq8v\" (UniqueName: \"kubernetes.io/projected/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-kube-api-access-glq8v\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.292054 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ceph\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.393672 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ssh-key\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.393807 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-inventory\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.393842 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glq8v\" (UniqueName: \"kubernetes.io/projected/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-kube-api-access-glq8v\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.393908 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ceph\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.398340 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ceph\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.398391 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ssh-key\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.399796 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-inventory\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.416736 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glq8v\" (UniqueName: \"kubernetes.io/projected/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-kube-api-access-glq8v\") pod \"configure-os-openstack-openstack-cell1-dv8tn\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:17 crc kubenswrapper[4770]: I1004 05:14:17.548157 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:14:18 crc kubenswrapper[4770]: I1004 05:14:18.049750 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-dv8tn"] Oct 04 05:14:18 crc kubenswrapper[4770]: I1004 05:14:18.158574 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" event={"ID":"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e","Type":"ContainerStarted","Data":"324570f975c5fe9546aa260cd5a05b2a9f44f9d467c4a1fa32d974a7dae6af86"} Oct 04 05:14:20 crc kubenswrapper[4770]: I1004 05:14:20.176758 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" event={"ID":"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e","Type":"ContainerStarted","Data":"3578adfcf7f45f0562d2d2f39b388daf035f8d9989ac38319a93c383fb213989"} Oct 04 05:14:20 crc kubenswrapper[4770]: I1004 05:14:20.199304 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" podStartSLOduration=2.328715198 podStartE2EDuration="3.199287596s" podCreationTimestamp="2025-10-04 05:14:17 +0000 UTC" firstStartedPulling="2025-10-04 05:14:18.056983016 +0000 UTC m=+7869.348992728" lastFinishedPulling="2025-10-04 05:14:18.927555414 +0000 UTC m=+7870.219565126" observedRunningTime="2025-10-04 05:14:20.195472506 +0000 UTC m=+7871.487482218" watchObservedRunningTime="2025-10-04 05:14:20.199287596 +0000 UTC m=+7871.491297308" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.185859 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz"] Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.188210 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.193461 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.193621 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.210733 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz"] Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.269902 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-config-volume\") pod \"collect-profiles-29325915-s6wnz\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.269986 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwqvt\" (UniqueName: \"kubernetes.io/projected/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-kube-api-access-gwqvt\") pod \"collect-profiles-29325915-s6wnz\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.270034 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-secret-volume\") pod \"collect-profiles-29325915-s6wnz\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.372090 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-config-volume\") pod \"collect-profiles-29325915-s6wnz\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.372137 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-secret-volume\") pod \"collect-profiles-29325915-s6wnz\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.372163 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwqvt\" (UniqueName: \"kubernetes.io/projected/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-kube-api-access-gwqvt\") pod \"collect-profiles-29325915-s6wnz\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.372993 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-config-volume\") pod \"collect-profiles-29325915-s6wnz\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.378666 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-secret-volume\") pod \"collect-profiles-29325915-s6wnz\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.391452 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwqvt\" (UniqueName: \"kubernetes.io/projected/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-kube-api-access-gwqvt\") pod \"collect-profiles-29325915-s6wnz\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:00 crc kubenswrapper[4770]: I1004 05:15:00.520077 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:01 crc kubenswrapper[4770]: I1004 05:15:01.046496 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz"] Oct 04 05:15:01 crc kubenswrapper[4770]: I1004 05:15:01.622578 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" event={"ID":"fc610f59-e16b-48b2-aef3-fee7f3a76cb1","Type":"ContainerStarted","Data":"90cc942168d13b57d964b1830177602a5b512b28713b93edf5b6ac523bf0aa40"} Oct 04 05:15:01 crc kubenswrapper[4770]: I1004 05:15:01.622905 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" event={"ID":"fc610f59-e16b-48b2-aef3-fee7f3a76cb1","Type":"ContainerStarted","Data":"ebaaa95bb9b20a274e80c1022417ce144d71c477879e0a2452e30a4b69d3dbe2"} Oct 04 05:15:02 crc kubenswrapper[4770]: I1004 05:15:02.637254 4770 generic.go:334] "Generic (PLEG): container finished" podID="fc610f59-e16b-48b2-aef3-fee7f3a76cb1" containerID="90cc942168d13b57d964b1830177602a5b512b28713b93edf5b6ac523bf0aa40" exitCode=0 Oct 04 05:15:02 crc kubenswrapper[4770]: I1004 05:15:02.637347 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" event={"ID":"fc610f59-e16b-48b2-aef3-fee7f3a76cb1","Type":"ContainerDied","Data":"90cc942168d13b57d964b1830177602a5b512b28713b93edf5b6ac523bf0aa40"} Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.048349 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.148822 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-secret-volume\") pod \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.148980 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-config-volume\") pod \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.149540 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwqvt\" (UniqueName: \"kubernetes.io/projected/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-kube-api-access-gwqvt\") pod \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\" (UID: \"fc610f59-e16b-48b2-aef3-fee7f3a76cb1\") " Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.149905 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-config-volume" (OuterVolumeSpecName: "config-volume") pod "fc610f59-e16b-48b2-aef3-fee7f3a76cb1" (UID: "fc610f59-e16b-48b2-aef3-fee7f3a76cb1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.150324 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.155112 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fc610f59-e16b-48b2-aef3-fee7f3a76cb1" (UID: "fc610f59-e16b-48b2-aef3-fee7f3a76cb1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.164375 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-kube-api-access-gwqvt" (OuterVolumeSpecName: "kube-api-access-gwqvt") pod "fc610f59-e16b-48b2-aef3-fee7f3a76cb1" (UID: "fc610f59-e16b-48b2-aef3-fee7f3a76cb1"). InnerVolumeSpecName "kube-api-access-gwqvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.252901 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.252948 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwqvt\" (UniqueName: \"kubernetes.io/projected/fc610f59-e16b-48b2-aef3-fee7f3a76cb1-kube-api-access-gwqvt\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.660659 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" event={"ID":"fc610f59-e16b-48b2-aef3-fee7f3a76cb1","Type":"ContainerDied","Data":"ebaaa95bb9b20a274e80c1022417ce144d71c477879e0a2452e30a4b69d3dbe2"} Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.661054 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebaaa95bb9b20a274e80c1022417ce144d71c477879e0a2452e30a4b69d3dbe2" Oct 04 05:15:04 crc kubenswrapper[4770]: I1004 05:15:04.660758 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz" Oct 04 05:15:05 crc kubenswrapper[4770]: I1004 05:15:05.138326 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7"] Oct 04 05:15:05 crc kubenswrapper[4770]: I1004 05:15:05.147520 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-6j6t7"] Oct 04 05:15:05 crc kubenswrapper[4770]: I1004 05:15:05.671364 4770 generic.go:334] "Generic (PLEG): container finished" podID="1ceac293-7f9f-45c4-a92a-8c3e82c4c85e" containerID="3578adfcf7f45f0562d2d2f39b388daf035f8d9989ac38319a93c383fb213989" exitCode=0 Oct 04 05:15:05 crc kubenswrapper[4770]: I1004 05:15:05.671407 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" event={"ID":"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e","Type":"ContainerDied","Data":"3578adfcf7f45f0562d2d2f39b388daf035f8d9989ac38319a93c383fb213989"} Oct 04 05:15:05 crc kubenswrapper[4770]: I1004 05:15:05.690126 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b51cff2d-a9f1-4262-9592-c4fc36dc299f" path="/var/lib/kubelet/pods/b51cff2d-a9f1-4262-9592-c4fc36dc299f/volumes" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.146503 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.217726 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ceph\") pod \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.217791 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ssh-key\") pod \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.217855 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-inventory\") pod \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.218040 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glq8v\" (UniqueName: \"kubernetes.io/projected/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-kube-api-access-glq8v\") pod \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\" (UID: \"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e\") " Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.225039 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ceph" (OuterVolumeSpecName: "ceph") pod "1ceac293-7f9f-45c4-a92a-8c3e82c4c85e" (UID: "1ceac293-7f9f-45c4-a92a-8c3e82c4c85e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.227235 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-kube-api-access-glq8v" (OuterVolumeSpecName: "kube-api-access-glq8v") pod "1ceac293-7f9f-45c4-a92a-8c3e82c4c85e" (UID: "1ceac293-7f9f-45c4-a92a-8c3e82c4c85e"). InnerVolumeSpecName "kube-api-access-glq8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.250959 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-inventory" (OuterVolumeSpecName: "inventory") pod "1ceac293-7f9f-45c4-a92a-8c3e82c4c85e" (UID: "1ceac293-7f9f-45c4-a92a-8c3e82c4c85e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.253329 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1ceac293-7f9f-45c4-a92a-8c3e82c4c85e" (UID: "1ceac293-7f9f-45c4-a92a-8c3e82c4c85e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.320614 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.320664 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.320681 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.320694 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glq8v\" (UniqueName: \"kubernetes.io/projected/1ceac293-7f9f-45c4-a92a-8c3e82c4c85e-kube-api-access-glq8v\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.693484 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" event={"ID":"1ceac293-7f9f-45c4-a92a-8c3e82c4c85e","Type":"ContainerDied","Data":"324570f975c5fe9546aa260cd5a05b2a9f44f9d467c4a1fa32d974a7dae6af86"} Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.693737 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="324570f975c5fe9546aa260cd5a05b2a9f44f9d467c4a1fa32d974a7dae6af86" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.693522 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-dv8tn" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.770897 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-rz6bz"] Oct 04 05:15:07 crc kubenswrapper[4770]: E1004 05:15:07.771328 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ceac293-7f9f-45c4-a92a-8c3e82c4c85e" containerName="configure-os-openstack-openstack-cell1" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.771346 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ceac293-7f9f-45c4-a92a-8c3e82c4c85e" containerName="configure-os-openstack-openstack-cell1" Oct 04 05:15:07 crc kubenswrapper[4770]: E1004 05:15:07.771377 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc610f59-e16b-48b2-aef3-fee7f3a76cb1" containerName="collect-profiles" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.771384 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc610f59-e16b-48b2-aef3-fee7f3a76cb1" containerName="collect-profiles" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.771633 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ceac293-7f9f-45c4-a92a-8c3e82c4c85e" containerName="configure-os-openstack-openstack-cell1" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.771670 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc610f59-e16b-48b2-aef3-fee7f3a76cb1" containerName="collect-profiles" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.772696 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.775313 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.775824 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.775965 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.777043 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.790762 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-rz6bz"] Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.830734 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.830880 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtv59\" (UniqueName: \"kubernetes.io/projected/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-kube-api-access-mtv59\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.830966 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ceph\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.831143 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-inventory-0\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.933403 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtv59\" (UniqueName: \"kubernetes.io/projected/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-kube-api-access-mtv59\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.933660 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ceph\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.934862 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-inventory-0\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.934993 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.939366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-inventory-0\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.940129 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ceph\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.942806 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:07 crc kubenswrapper[4770]: I1004 05:15:07.952357 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtv59\" (UniqueName: \"kubernetes.io/projected/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-kube-api-access-mtv59\") pod \"ssh-known-hosts-openstack-rz6bz\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:08 crc kubenswrapper[4770]: I1004 05:15:08.096445 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:08 crc kubenswrapper[4770]: I1004 05:15:08.713445 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-rz6bz"] Oct 04 05:15:09 crc kubenswrapper[4770]: I1004 05:15:09.719883 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rz6bz" event={"ID":"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1","Type":"ContainerStarted","Data":"b4be4f6490fceb07a72832c500241bc715ef68c21a09720dfcc13cca650387e5"} Oct 04 05:15:10 crc kubenswrapper[4770]: I1004 05:15:10.734037 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rz6bz" event={"ID":"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1","Type":"ContainerStarted","Data":"85d3aa5ac876c6f1243a88c64a5b15bf9338f29b32e246579842da18dc89eb9a"} Oct 04 05:15:10 crc kubenswrapper[4770]: I1004 05:15:10.758371 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-rz6bz" podStartSLOduration=2.896916869 podStartE2EDuration="3.758349489s" podCreationTimestamp="2025-10-04 05:15:07 +0000 UTC" firstStartedPulling="2025-10-04 05:15:08.711385703 +0000 UTC m=+7920.003395405" lastFinishedPulling="2025-10-04 05:15:09.572818303 +0000 UTC m=+7920.864828025" observedRunningTime="2025-10-04 05:15:10.747543686 +0000 UTC m=+7922.039553418" watchObservedRunningTime="2025-10-04 05:15:10.758349489 +0000 UTC m=+7922.050359211" Oct 04 05:15:19 crc kubenswrapper[4770]: I1004 05:15:19.826718 4770 generic.go:334] "Generic (PLEG): container finished" podID="13e4633f-4f16-4e0c-8cfb-b500a9ba63f1" containerID="85d3aa5ac876c6f1243a88c64a5b15bf9338f29b32e246579842da18dc89eb9a" exitCode=0 Oct 04 05:15:19 crc kubenswrapper[4770]: I1004 05:15:19.826817 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rz6bz" event={"ID":"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1","Type":"ContainerDied","Data":"85d3aa5ac876c6f1243a88c64a5b15bf9338f29b32e246579842da18dc89eb9a"} Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.570121 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.645918 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ssh-key-openstack-cell1\") pod \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.645993 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-inventory-0\") pod \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.646046 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ceph\") pod \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.646122 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtv59\" (UniqueName: \"kubernetes.io/projected/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-kube-api-access-mtv59\") pod \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\" (UID: \"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1\") " Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.651877 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ceph" (OuterVolumeSpecName: "ceph") pod "13e4633f-4f16-4e0c-8cfb-b500a9ba63f1" (UID: "13e4633f-4f16-4e0c-8cfb-b500a9ba63f1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.652985 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-kube-api-access-mtv59" (OuterVolumeSpecName: "kube-api-access-mtv59") pod "13e4633f-4f16-4e0c-8cfb-b500a9ba63f1" (UID: "13e4633f-4f16-4e0c-8cfb-b500a9ba63f1"). InnerVolumeSpecName "kube-api-access-mtv59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.684232 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "13e4633f-4f16-4e0c-8cfb-b500a9ba63f1" (UID: "13e4633f-4f16-4e0c-8cfb-b500a9ba63f1"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.687429 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "13e4633f-4f16-4e0c-8cfb-b500a9ba63f1" (UID: "13e4633f-4f16-4e0c-8cfb-b500a9ba63f1"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.748617 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtv59\" (UniqueName: \"kubernetes.io/projected/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-kube-api-access-mtv59\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.748657 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.748696 4770 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.748709 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/13e4633f-4f16-4e0c-8cfb-b500a9ba63f1-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.852202 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rz6bz" event={"ID":"13e4633f-4f16-4e0c-8cfb-b500a9ba63f1","Type":"ContainerDied","Data":"b4be4f6490fceb07a72832c500241bc715ef68c21a09720dfcc13cca650387e5"} Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.852282 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4be4f6490fceb07a72832c500241bc715ef68c21a09720dfcc13cca650387e5" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.852223 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rz6bz" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.908532 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-p7q9s"] Oct 04 05:15:21 crc kubenswrapper[4770]: E1004 05:15:21.908992 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e4633f-4f16-4e0c-8cfb-b500a9ba63f1" containerName="ssh-known-hosts-openstack" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.909009 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e4633f-4f16-4e0c-8cfb-b500a9ba63f1" containerName="ssh-known-hosts-openstack" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.909240 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="13e4633f-4f16-4e0c-8cfb-b500a9ba63f1" containerName="ssh-known-hosts-openstack" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.910682 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.913051 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.913612 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.913646 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.913878 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.924356 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-p7q9s"] Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.953093 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ssh-key\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.953382 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-inventory\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.953493 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ceph\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:21 crc kubenswrapper[4770]: I1004 05:15:21.953854 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgmt6\" (UniqueName: \"kubernetes.io/projected/a8ebb9de-f667-40ba-9d22-897cbabb320e-kube-api-access-hgmt6\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.056412 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ssh-key\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.056557 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-inventory\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.056596 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ceph\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.056633 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgmt6\" (UniqueName: \"kubernetes.io/projected/a8ebb9de-f667-40ba-9d22-897cbabb320e-kube-api-access-hgmt6\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.061299 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-inventory\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.063960 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ssh-key\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.074404 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgmt6\" (UniqueName: \"kubernetes.io/projected/a8ebb9de-f667-40ba-9d22-897cbabb320e-kube-api-access-hgmt6\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.078159 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ceph\") pod \"run-os-openstack-openstack-cell1-p7q9s\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.231183 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.759408 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-p7q9s"] Oct 04 05:15:22 crc kubenswrapper[4770]: I1004 05:15:22.865415 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-p7q9s" event={"ID":"a8ebb9de-f667-40ba-9d22-897cbabb320e","Type":"ContainerStarted","Data":"c17525599d3afc718a398d95d5312b2b636c757ad26bc6f6d4ddb9d03c98092b"} Oct 04 05:15:23 crc kubenswrapper[4770]: I1004 05:15:23.880435 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-p7q9s" event={"ID":"a8ebb9de-f667-40ba-9d22-897cbabb320e","Type":"ContainerStarted","Data":"26fa55a4fe7afcd8c0cf11c8c3705f9a09abaad215cfae9420438b0125cce755"} Oct 04 05:15:23 crc kubenswrapper[4770]: I1004 05:15:23.915671 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-p7q9s" podStartSLOduration=2.340180816 podStartE2EDuration="2.915646207s" podCreationTimestamp="2025-10-04 05:15:21 +0000 UTC" firstStartedPulling="2025-10-04 05:15:22.764346786 +0000 UTC m=+7934.056356498" lastFinishedPulling="2025-10-04 05:15:23.339812177 +0000 UTC m=+7934.631821889" observedRunningTime="2025-10-04 05:15:23.902311878 +0000 UTC m=+7935.194321590" watchObservedRunningTime="2025-10-04 05:15:23.915646207 +0000 UTC m=+7935.207655919" Oct 04 05:15:31 crc kubenswrapper[4770]: I1004 05:15:31.407637 4770 scope.go:117] "RemoveContainer" containerID="cd6c1de9e081ea52546c1269f24f0471edf2fcd487697b6a21c172d5a68941c5" Oct 04 05:15:31 crc kubenswrapper[4770]: I1004 05:15:31.959882 4770 generic.go:334] "Generic (PLEG): container finished" podID="a8ebb9de-f667-40ba-9d22-897cbabb320e" containerID="26fa55a4fe7afcd8c0cf11c8c3705f9a09abaad215cfae9420438b0125cce755" exitCode=0 Oct 04 05:15:31 crc kubenswrapper[4770]: I1004 05:15:31.959975 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-p7q9s" event={"ID":"a8ebb9de-f667-40ba-9d22-897cbabb320e","Type":"ContainerDied","Data":"26fa55a4fe7afcd8c0cf11c8c3705f9a09abaad215cfae9420438b0125cce755"} Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.466847 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.610799 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgmt6\" (UniqueName: \"kubernetes.io/projected/a8ebb9de-f667-40ba-9d22-897cbabb320e-kube-api-access-hgmt6\") pod \"a8ebb9de-f667-40ba-9d22-897cbabb320e\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.610928 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ssh-key\") pod \"a8ebb9de-f667-40ba-9d22-897cbabb320e\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.610996 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-inventory\") pod \"a8ebb9de-f667-40ba-9d22-897cbabb320e\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.611028 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ceph\") pod \"a8ebb9de-f667-40ba-9d22-897cbabb320e\" (UID: \"a8ebb9de-f667-40ba-9d22-897cbabb320e\") " Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.616835 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ceph" (OuterVolumeSpecName: "ceph") pod "a8ebb9de-f667-40ba-9d22-897cbabb320e" (UID: "a8ebb9de-f667-40ba-9d22-897cbabb320e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.633506 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8ebb9de-f667-40ba-9d22-897cbabb320e-kube-api-access-hgmt6" (OuterVolumeSpecName: "kube-api-access-hgmt6") pod "a8ebb9de-f667-40ba-9d22-897cbabb320e" (UID: "a8ebb9de-f667-40ba-9d22-897cbabb320e"). InnerVolumeSpecName "kube-api-access-hgmt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.641890 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a8ebb9de-f667-40ba-9d22-897cbabb320e" (UID: "a8ebb9de-f667-40ba-9d22-897cbabb320e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.655860 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-inventory" (OuterVolumeSpecName: "inventory") pod "a8ebb9de-f667-40ba-9d22-897cbabb320e" (UID: "a8ebb9de-f667-40ba-9d22-897cbabb320e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.716684 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgmt6\" (UniqueName: \"kubernetes.io/projected/a8ebb9de-f667-40ba-9d22-897cbabb320e-kube-api-access-hgmt6\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.716720 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.716732 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.716743 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8ebb9de-f667-40ba-9d22-897cbabb320e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.979998 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-p7q9s" event={"ID":"a8ebb9de-f667-40ba-9d22-897cbabb320e","Type":"ContainerDied","Data":"c17525599d3afc718a398d95d5312b2b636c757ad26bc6f6d4ddb9d03c98092b"} Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.980371 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c17525599d3afc718a398d95d5312b2b636c757ad26bc6f6d4ddb9d03c98092b" Oct 04 05:15:33 crc kubenswrapper[4770]: I1004 05:15:33.980092 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-p7q9s" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.093884 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-2v52b"] Oct 04 05:15:34 crc kubenswrapper[4770]: E1004 05:15:34.094388 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8ebb9de-f667-40ba-9d22-897cbabb320e" containerName="run-os-openstack-openstack-cell1" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.094408 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8ebb9de-f667-40ba-9d22-897cbabb320e" containerName="run-os-openstack-openstack-cell1" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.094664 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8ebb9de-f667-40ba-9d22-897cbabb320e" containerName="run-os-openstack-openstack-cell1" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.095602 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.099733 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.100281 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.100314 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.101535 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.107483 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-2v52b"] Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.227610 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-inventory\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.228028 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.228259 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbvkt\" (UniqueName: \"kubernetes.io/projected/704ceb18-dc32-4a0d-ad0a-dee111bae162-kube-api-access-rbvkt\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.228384 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ceph\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.330628 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ceph\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.330733 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-inventory\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.330877 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.331059 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbvkt\" (UniqueName: \"kubernetes.io/projected/704ceb18-dc32-4a0d-ad0a-dee111bae162-kube-api-access-rbvkt\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.334668 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-inventory\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.335053 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.349325 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ceph\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.352820 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbvkt\" (UniqueName: \"kubernetes.io/projected/704ceb18-dc32-4a0d-ad0a-dee111bae162-kube-api-access-rbvkt\") pod \"reboot-os-openstack-openstack-cell1-2v52b\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:34 crc kubenswrapper[4770]: I1004 05:15:34.426591 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:37 crc kubenswrapper[4770]: I1004 05:15:35.164321 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-2v52b"] Oct 04 05:15:37 crc kubenswrapper[4770]: I1004 05:15:35.999361 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" event={"ID":"704ceb18-dc32-4a0d-ad0a-dee111bae162","Type":"ContainerStarted","Data":"c9b829a581a4ff66f62ec7db45df16584d9a4abc7f862faf1d9257518f04c715"} Oct 04 05:15:37 crc kubenswrapper[4770]: I1004 05:15:37.011702 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" event={"ID":"704ceb18-dc32-4a0d-ad0a-dee111bae162","Type":"ContainerStarted","Data":"b258b9c95d025877e8eb74892abe1a79b443d83da95ca65d4de1170478511816"} Oct 04 05:15:37 crc kubenswrapper[4770]: I1004 05:15:37.035830 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" podStartSLOduration=1.741762856 podStartE2EDuration="3.035812571s" podCreationTimestamp="2025-10-04 05:15:34 +0000 UTC" firstStartedPulling="2025-10-04 05:15:35.170874116 +0000 UTC m=+7946.462883828" lastFinishedPulling="2025-10-04 05:15:36.464923841 +0000 UTC m=+7947.756933543" observedRunningTime="2025-10-04 05:15:37.0330817 +0000 UTC m=+7948.325091432" watchObservedRunningTime="2025-10-04 05:15:37.035812571 +0000 UTC m=+7948.327822293" Oct 04 05:15:43 crc kubenswrapper[4770]: I1004 05:15:43.945849 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bs6kx"] Oct 04 05:15:43 crc kubenswrapper[4770]: I1004 05:15:43.955198 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:43 crc kubenswrapper[4770]: I1004 05:15:43.961946 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bs6kx"] Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.055544 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-catalog-content\") pod \"certified-operators-bs6kx\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.056197 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqpbf\" (UniqueName: \"kubernetes.io/projected/14571d79-5289-48a5-b7c0-49d0e1ba0690-kube-api-access-gqpbf\") pod \"certified-operators-bs6kx\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.056997 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-utilities\") pod \"certified-operators-bs6kx\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.163739 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-utilities\") pod \"certified-operators-bs6kx\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.164194 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-catalog-content\") pod \"certified-operators-bs6kx\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.164298 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqpbf\" (UniqueName: \"kubernetes.io/projected/14571d79-5289-48a5-b7c0-49d0e1ba0690-kube-api-access-gqpbf\") pod \"certified-operators-bs6kx\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.164470 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-utilities\") pod \"certified-operators-bs6kx\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.164950 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-catalog-content\") pod \"certified-operators-bs6kx\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.192075 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqpbf\" (UniqueName: \"kubernetes.io/projected/14571d79-5289-48a5-b7c0-49d0e1ba0690-kube-api-access-gqpbf\") pod \"certified-operators-bs6kx\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.311823 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:15:44 crc kubenswrapper[4770]: I1004 05:15:44.867397 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bs6kx"] Oct 04 05:15:45 crc kubenswrapper[4770]: I1004 05:15:45.099321 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bs6kx" event={"ID":"14571d79-5289-48a5-b7c0-49d0e1ba0690","Type":"ContainerStarted","Data":"bab52752be8bdb0e5fbf49bc9b47ccbc1df5c392a3ffeb7f7f3f727e804d8a3e"} Oct 04 05:15:46 crc kubenswrapper[4770]: I1004 05:15:46.115048 4770 generic.go:334] "Generic (PLEG): container finished" podID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerID="3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825" exitCode=0 Oct 04 05:15:46 crc kubenswrapper[4770]: I1004 05:15:46.115387 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bs6kx" event={"ID":"14571d79-5289-48a5-b7c0-49d0e1ba0690","Type":"ContainerDied","Data":"3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825"} Oct 04 05:15:48 crc kubenswrapper[4770]: I1004 05:15:48.137326 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bs6kx" event={"ID":"14571d79-5289-48a5-b7c0-49d0e1ba0690","Type":"ContainerStarted","Data":"6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2"} Oct 04 05:15:55 crc kubenswrapper[4770]: I1004 05:15:55.219076 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" event={"ID":"704ceb18-dc32-4a0d-ad0a-dee111bae162","Type":"ContainerDied","Data":"b258b9c95d025877e8eb74892abe1a79b443d83da95ca65d4de1170478511816"} Oct 04 05:15:55 crc kubenswrapper[4770]: I1004 05:15:55.218988 4770 generic.go:334] "Generic (PLEG): container finished" podID="704ceb18-dc32-4a0d-ad0a-dee111bae162" containerID="b258b9c95d025877e8eb74892abe1a79b443d83da95ca65d4de1170478511816" exitCode=0 Oct 04 05:15:55 crc kubenswrapper[4770]: I1004 05:15:55.224563 4770 generic.go:334] "Generic (PLEG): container finished" podID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerID="6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2" exitCode=0 Oct 04 05:15:55 crc kubenswrapper[4770]: I1004 05:15:55.224683 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bs6kx" event={"ID":"14571d79-5289-48a5-b7c0-49d0e1ba0690","Type":"ContainerDied","Data":"6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2"} Oct 04 05:15:55 crc kubenswrapper[4770]: I1004 05:15:55.229651 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:15:56 crc kubenswrapper[4770]: I1004 05:15:56.873079 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.027977 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-inventory\") pod \"704ceb18-dc32-4a0d-ad0a-dee111bae162\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.028058 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ceph\") pod \"704ceb18-dc32-4a0d-ad0a-dee111bae162\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.028235 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbvkt\" (UniqueName: \"kubernetes.io/projected/704ceb18-dc32-4a0d-ad0a-dee111bae162-kube-api-access-rbvkt\") pod \"704ceb18-dc32-4a0d-ad0a-dee111bae162\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.028267 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ssh-key\") pod \"704ceb18-dc32-4a0d-ad0a-dee111bae162\" (UID: \"704ceb18-dc32-4a0d-ad0a-dee111bae162\") " Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.034003 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/704ceb18-dc32-4a0d-ad0a-dee111bae162-kube-api-access-rbvkt" (OuterVolumeSpecName: "kube-api-access-rbvkt") pod "704ceb18-dc32-4a0d-ad0a-dee111bae162" (UID: "704ceb18-dc32-4a0d-ad0a-dee111bae162"). InnerVolumeSpecName "kube-api-access-rbvkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.034856 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ceph" (OuterVolumeSpecName: "ceph") pod "704ceb18-dc32-4a0d-ad0a-dee111bae162" (UID: "704ceb18-dc32-4a0d-ad0a-dee111bae162"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.064938 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-inventory" (OuterVolumeSpecName: "inventory") pod "704ceb18-dc32-4a0d-ad0a-dee111bae162" (UID: "704ceb18-dc32-4a0d-ad0a-dee111bae162"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.068320 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "704ceb18-dc32-4a0d-ad0a-dee111bae162" (UID: "704ceb18-dc32-4a0d-ad0a-dee111bae162"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.132408 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbvkt\" (UniqueName: \"kubernetes.io/projected/704ceb18-dc32-4a0d-ad0a-dee111bae162-kube-api-access-rbvkt\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.132448 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.132458 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.132466 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/704ceb18-dc32-4a0d-ad0a-dee111bae162-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.253263 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bs6kx" event={"ID":"14571d79-5289-48a5-b7c0-49d0e1ba0690","Type":"ContainerStarted","Data":"068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342"} Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.255965 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" event={"ID":"704ceb18-dc32-4a0d-ad0a-dee111bae162","Type":"ContainerDied","Data":"c9b829a581a4ff66f62ec7db45df16584d9a4abc7f862faf1d9257518f04c715"} Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.255998 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9b829a581a4ff66f62ec7db45df16584d9a4abc7f862faf1d9257518f04c715" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.256055 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-2v52b" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.287535 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bs6kx" podStartSLOduration=4.206455364 podStartE2EDuration="14.287517225s" podCreationTimestamp="2025-10-04 05:15:43 +0000 UTC" firstStartedPulling="2025-10-04 05:15:46.117673909 +0000 UTC m=+7957.409683611" lastFinishedPulling="2025-10-04 05:15:56.19873575 +0000 UTC m=+7967.490745472" observedRunningTime="2025-10-04 05:15:57.28653267 +0000 UTC m=+7968.578542412" watchObservedRunningTime="2025-10-04 05:15:57.287517225 +0000 UTC m=+7968.579526937" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.337266 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-vklws"] Oct 04 05:15:57 crc kubenswrapper[4770]: E1004 05:15:57.337768 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="704ceb18-dc32-4a0d-ad0a-dee111bae162" containerName="reboot-os-openstack-openstack-cell1" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.337788 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="704ceb18-dc32-4a0d-ad0a-dee111bae162" containerName="reboot-os-openstack-openstack-cell1" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.338113 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="704ceb18-dc32-4a0d-ad0a-dee111bae162" containerName="reboot-os-openstack-openstack-cell1" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.339471 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.347401 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.347577 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.347715 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.347880 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.350620 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-vklws"] Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.439280 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ceph\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.439494 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh7fp\" (UniqueName: \"kubernetes.io/projected/3cffb45c-2e95-4e28-9c87-56593c1611df-kube-api-access-gh7fp\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.439518 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.439559 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.439665 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-inventory\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.439719 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.439746 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.439872 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.441335 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.441481 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.441546 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.441566 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ssh-key\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.543802 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.543875 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.543906 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.543930 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ssh-key\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.544019 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ceph\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.544117 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh7fp\" (UniqueName: \"kubernetes.io/projected/3cffb45c-2e95-4e28-9c87-56593c1611df-kube-api-access-gh7fp\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.544143 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.544175 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.544232 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-inventory\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.544264 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.544298 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.544324 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.550934 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.551422 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.551562 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.551588 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ssh-key\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.551697 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.552502 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.552837 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.553069 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.553626 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ceph\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.553878 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.556709 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-inventory\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.571657 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh7fp\" (UniqueName: \"kubernetes.io/projected/3cffb45c-2e95-4e28-9c87-56593c1611df-kube-api-access-gh7fp\") pod \"install-certs-openstack-openstack-cell1-vklws\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:57 crc kubenswrapper[4770]: I1004 05:15:57.676468 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:15:58 crc kubenswrapper[4770]: I1004 05:15:58.294368 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-vklws"] Oct 04 05:15:58 crc kubenswrapper[4770]: W1004 05:15:58.297945 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cffb45c_2e95_4e28_9c87_56593c1611df.slice/crio-51038096fd0dabacdb964c6b0a54514ab8429c22effdd23a9280c2f29fce5bfd WatchSource:0}: Error finding container 51038096fd0dabacdb964c6b0a54514ab8429c22effdd23a9280c2f29fce5bfd: Status 404 returned error can't find the container with id 51038096fd0dabacdb964c6b0a54514ab8429c22effdd23a9280c2f29fce5bfd Oct 04 05:15:59 crc kubenswrapper[4770]: I1004 05:15:59.280617 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vklws" event={"ID":"3cffb45c-2e95-4e28-9c87-56593c1611df","Type":"ContainerStarted","Data":"51038096fd0dabacdb964c6b0a54514ab8429c22effdd23a9280c2f29fce5bfd"} Oct 04 05:16:00 crc kubenswrapper[4770]: I1004 05:16:00.292397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vklws" event={"ID":"3cffb45c-2e95-4e28-9c87-56593c1611df","Type":"ContainerStarted","Data":"38389b7933b3db82247b9db02c6f8b4f5ffe5c519e45b1edfc6c9fdd39bfcd9a"} Oct 04 05:16:00 crc kubenswrapper[4770]: I1004 05:16:00.313306 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-vklws" podStartSLOduration=2.292840623 podStartE2EDuration="3.313285462s" podCreationTimestamp="2025-10-04 05:15:57 +0000 UTC" firstStartedPulling="2025-10-04 05:15:58.301065644 +0000 UTC m=+7969.593075366" lastFinishedPulling="2025-10-04 05:15:59.321510483 +0000 UTC m=+7970.613520205" observedRunningTime="2025-10-04 05:16:00.309269158 +0000 UTC m=+7971.601278890" watchObservedRunningTime="2025-10-04 05:16:00.313285462 +0000 UTC m=+7971.605295184" Oct 04 05:16:01 crc kubenswrapper[4770]: I1004 05:16:01.795723 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:16:01 crc kubenswrapper[4770]: I1004 05:16:01.796305 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:16:04 crc kubenswrapper[4770]: I1004 05:16:04.312246 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:16:04 crc kubenswrapper[4770]: I1004 05:16:04.312670 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:16:04 crc kubenswrapper[4770]: I1004 05:16:04.361830 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:16:04 crc kubenswrapper[4770]: I1004 05:16:04.415460 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:16:04 crc kubenswrapper[4770]: I1004 05:16:04.598424 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bs6kx"] Oct 04 05:16:06 crc kubenswrapper[4770]: I1004 05:16:06.353708 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bs6kx" podUID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerName="registry-server" containerID="cri-o://068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342" gracePeriod=2 Oct 04 05:16:06 crc kubenswrapper[4770]: I1004 05:16:06.961535 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.060854 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqpbf\" (UniqueName: \"kubernetes.io/projected/14571d79-5289-48a5-b7c0-49d0e1ba0690-kube-api-access-gqpbf\") pod \"14571d79-5289-48a5-b7c0-49d0e1ba0690\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.060932 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-catalog-content\") pod \"14571d79-5289-48a5-b7c0-49d0e1ba0690\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.061053 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-utilities\") pod \"14571d79-5289-48a5-b7c0-49d0e1ba0690\" (UID: \"14571d79-5289-48a5-b7c0-49d0e1ba0690\") " Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.061861 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-utilities" (OuterVolumeSpecName: "utilities") pod "14571d79-5289-48a5-b7c0-49d0e1ba0690" (UID: "14571d79-5289-48a5-b7c0-49d0e1ba0690"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.066268 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14571d79-5289-48a5-b7c0-49d0e1ba0690-kube-api-access-gqpbf" (OuterVolumeSpecName: "kube-api-access-gqpbf") pod "14571d79-5289-48a5-b7c0-49d0e1ba0690" (UID: "14571d79-5289-48a5-b7c0-49d0e1ba0690"). InnerVolumeSpecName "kube-api-access-gqpbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.121808 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14571d79-5289-48a5-b7c0-49d0e1ba0690" (UID: "14571d79-5289-48a5-b7c0-49d0e1ba0690"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.165375 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.165447 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14571d79-5289-48a5-b7c0-49d0e1ba0690-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.165462 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqpbf\" (UniqueName: \"kubernetes.io/projected/14571d79-5289-48a5-b7c0-49d0e1ba0690-kube-api-access-gqpbf\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.364747 4770 generic.go:334] "Generic (PLEG): container finished" podID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerID="068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342" exitCode=0 Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.364810 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bs6kx" event={"ID":"14571d79-5289-48a5-b7c0-49d0e1ba0690","Type":"ContainerDied","Data":"068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342"} Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.364825 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bs6kx" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.366075 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bs6kx" event={"ID":"14571d79-5289-48a5-b7c0-49d0e1ba0690","Type":"ContainerDied","Data":"bab52752be8bdb0e5fbf49bc9b47ccbc1df5c392a3ffeb7f7f3f727e804d8a3e"} Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.366138 4770 scope.go:117] "RemoveContainer" containerID="068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.397847 4770 scope.go:117] "RemoveContainer" containerID="6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.420203 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bs6kx"] Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.431624 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bs6kx"] Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.455025 4770 scope.go:117] "RemoveContainer" containerID="3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.473246 4770 scope.go:117] "RemoveContainer" containerID="068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342" Oct 04 05:16:07 crc kubenswrapper[4770]: E1004 05:16:07.473817 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342\": container with ID starting with 068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342 not found: ID does not exist" containerID="068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.473889 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342"} err="failed to get container status \"068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342\": rpc error: code = NotFound desc = could not find container \"068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342\": container with ID starting with 068914827f408f7419901dd4e20dbc2d5df101395c9e7c2c4c65eab834502342 not found: ID does not exist" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.473936 4770 scope.go:117] "RemoveContainer" containerID="6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2" Oct 04 05:16:07 crc kubenswrapper[4770]: E1004 05:16:07.474302 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2\": container with ID starting with 6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2 not found: ID does not exist" containerID="6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.474330 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2"} err="failed to get container status \"6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2\": rpc error: code = NotFound desc = could not find container \"6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2\": container with ID starting with 6788aaa2f916ee57521604f824edb56060c2f59421206745fad0c9fb3a4351f2 not found: ID does not exist" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.474349 4770 scope.go:117] "RemoveContainer" containerID="3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825" Oct 04 05:16:07 crc kubenswrapper[4770]: E1004 05:16:07.474967 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825\": container with ID starting with 3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825 not found: ID does not exist" containerID="3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.475030 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825"} err="failed to get container status \"3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825\": rpc error: code = NotFound desc = could not find container \"3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825\": container with ID starting with 3df63b41bc44957e805b2d90d33c49d13a3c78bf86dd2a958246c7027eb01825 not found: ID does not exist" Oct 04 05:16:07 crc kubenswrapper[4770]: I1004 05:16:07.687715 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14571d79-5289-48a5-b7c0-49d0e1ba0690" path="/var/lib/kubelet/pods/14571d79-5289-48a5-b7c0-49d0e1ba0690/volumes" Oct 04 05:16:19 crc kubenswrapper[4770]: I1004 05:16:19.512454 4770 generic.go:334] "Generic (PLEG): container finished" podID="3cffb45c-2e95-4e28-9c87-56593c1611df" containerID="38389b7933b3db82247b9db02c6f8b4f5ffe5c519e45b1edfc6c9fdd39bfcd9a" exitCode=0 Oct 04 05:16:19 crc kubenswrapper[4770]: I1004 05:16:19.512563 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vklws" event={"ID":"3cffb45c-2e95-4e28-9c87-56593c1611df","Type":"ContainerDied","Data":"38389b7933b3db82247b9db02c6f8b4f5ffe5c519e45b1edfc6c9fdd39bfcd9a"} Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.053552 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.106406 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-libvirt-combined-ca-bundle\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.106544 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-dhcp-combined-ca-bundle\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.106640 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-nova-combined-ca-bundle\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.106712 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-inventory\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.106759 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-telemetry-combined-ca-bundle\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.106863 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh7fp\" (UniqueName: \"kubernetes.io/projected/3cffb45c-2e95-4e28-9c87-56593c1611df-kube-api-access-gh7fp\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.106905 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ceph\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.106971 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-sriov-combined-ca-bundle\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.107051 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-metadata-combined-ca-bundle\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.107089 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-bootstrap-combined-ca-bundle\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.107317 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ovn-combined-ca-bundle\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.107408 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ssh-key\") pod \"3cffb45c-2e95-4e28-9c87-56593c1611df\" (UID: \"3cffb45c-2e95-4e28-9c87-56593c1611df\") " Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.113629 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cffb45c-2e95-4e28-9c87-56593c1611df-kube-api-access-gh7fp" (OuterVolumeSpecName: "kube-api-access-gh7fp") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "kube-api-access-gh7fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.113986 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.114046 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.114868 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.115736 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.115780 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.115967 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ceph" (OuterVolumeSpecName: "ceph") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.116265 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.116573 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.124471 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.150926 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.152214 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-inventory" (OuterVolumeSpecName: "inventory") pod "3cffb45c-2e95-4e28-9c87-56593c1611df" (UID: "3cffb45c-2e95-4e28-9c87-56593c1611df"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210156 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh7fp\" (UniqueName: \"kubernetes.io/projected/3cffb45c-2e95-4e28-9c87-56593c1611df-kube-api-access-gh7fp\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210188 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210200 4770 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210210 4770 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210220 4770 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210229 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210236 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210244 4770 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210252 4770 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210261 4770 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210269 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.210278 4770 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cffb45c-2e95-4e28-9c87-56593c1611df-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.535831 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vklws" event={"ID":"3cffb45c-2e95-4e28-9c87-56593c1611df","Type":"ContainerDied","Data":"51038096fd0dabacdb964c6b0a54514ab8429c22effdd23a9280c2f29fce5bfd"} Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.536190 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51038096fd0dabacdb964c6b0a54514ab8429c22effdd23a9280c2f29fce5bfd" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.535906 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vklws" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.622403 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-4zjfm"] Oct 04 05:16:21 crc kubenswrapper[4770]: E1004 05:16:21.622937 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerName="extract-content" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.622959 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerName="extract-content" Oct 04 05:16:21 crc kubenswrapper[4770]: E1004 05:16:21.622986 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cffb45c-2e95-4e28-9c87-56593c1611df" containerName="install-certs-openstack-openstack-cell1" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.622998 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cffb45c-2e95-4e28-9c87-56593c1611df" containerName="install-certs-openstack-openstack-cell1" Oct 04 05:16:21 crc kubenswrapper[4770]: E1004 05:16:21.623053 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerName="registry-server" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.623062 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerName="registry-server" Oct 04 05:16:21 crc kubenswrapper[4770]: E1004 05:16:21.623076 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerName="extract-utilities" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.623083 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerName="extract-utilities" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.623343 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cffb45c-2e95-4e28-9c87-56593c1611df" containerName="install-certs-openstack-openstack-cell1" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.623368 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="14571d79-5289-48a5-b7c0-49d0e1ba0690" containerName="registry-server" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.624214 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.627183 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.627266 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.627477 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.627550 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.640169 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-4zjfm"] Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.721074 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t25sn\" (UniqueName: \"kubernetes.io/projected/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-kube-api-access-t25sn\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.721388 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-inventory\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.722606 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ceph\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.723230 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.828156 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t25sn\" (UniqueName: \"kubernetes.io/projected/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-kube-api-access-t25sn\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.828465 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-inventory\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.828674 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ceph\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.828830 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.833259 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-inventory\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.834243 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.835870 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ceph\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.846801 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t25sn\" (UniqueName: \"kubernetes.io/projected/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-kube-api-access-t25sn\") pod \"ceph-client-openstack-openstack-cell1-4zjfm\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:21 crc kubenswrapper[4770]: I1004 05:16:21.944144 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:22 crc kubenswrapper[4770]: I1004 05:16:22.574834 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-4zjfm"] Oct 04 05:16:22 crc kubenswrapper[4770]: W1004 05:16:22.586325 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod654841f2_f5ef_4b6a_bc0f_d78b8dbe5347.slice/crio-69746dbaed87227163fabcd995ef962fc049c533b3cf38ae3316e2ee22dffcfc WatchSource:0}: Error finding container 69746dbaed87227163fabcd995ef962fc049c533b3cf38ae3316e2ee22dffcfc: Status 404 returned error can't find the container with id 69746dbaed87227163fabcd995ef962fc049c533b3cf38ae3316e2ee22dffcfc Oct 04 05:16:23 crc kubenswrapper[4770]: I1004 05:16:23.559466 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" event={"ID":"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347","Type":"ContainerStarted","Data":"69746dbaed87227163fabcd995ef962fc049c533b3cf38ae3316e2ee22dffcfc"} Oct 04 05:16:25 crc kubenswrapper[4770]: I1004 05:16:25.583100 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" event={"ID":"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347","Type":"ContainerStarted","Data":"2efd1a54bbae1eb8cae9bb0fbde7072a891d2367af0a4102e234fc3e1b574fbc"} Oct 04 05:16:25 crc kubenswrapper[4770]: I1004 05:16:25.608568 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" podStartSLOduration=2.6559864380000002 podStartE2EDuration="4.608550795s" podCreationTimestamp="2025-10-04 05:16:21 +0000 UTC" firstStartedPulling="2025-10-04 05:16:22.58856403 +0000 UTC m=+7993.880573742" lastFinishedPulling="2025-10-04 05:16:24.541128387 +0000 UTC m=+7995.833138099" observedRunningTime="2025-10-04 05:16:25.597558937 +0000 UTC m=+7996.889568659" watchObservedRunningTime="2025-10-04 05:16:25.608550795 +0000 UTC m=+7996.900560507" Oct 04 05:16:29 crc kubenswrapper[4770]: I1004 05:16:29.631776 4770 generic.go:334] "Generic (PLEG): container finished" podID="654841f2-f5ef-4b6a-bc0f-d78b8dbe5347" containerID="2efd1a54bbae1eb8cae9bb0fbde7072a891d2367af0a4102e234fc3e1b574fbc" exitCode=0 Oct 04 05:16:29 crc kubenswrapper[4770]: I1004 05:16:29.631855 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" event={"ID":"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347","Type":"ContainerDied","Data":"2efd1a54bbae1eb8cae9bb0fbde7072a891d2367af0a4102e234fc3e1b574fbc"} Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.141763 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.242114 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-inventory\") pod \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.242581 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t25sn\" (UniqueName: \"kubernetes.io/projected/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-kube-api-access-t25sn\") pod \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.242750 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ssh-key\") pod \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.242920 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ceph\") pod \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\" (UID: \"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347\") " Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.250716 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-kube-api-access-t25sn" (OuterVolumeSpecName: "kube-api-access-t25sn") pod "654841f2-f5ef-4b6a-bc0f-d78b8dbe5347" (UID: "654841f2-f5ef-4b6a-bc0f-d78b8dbe5347"). InnerVolumeSpecName "kube-api-access-t25sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.252904 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ceph" (OuterVolumeSpecName: "ceph") pod "654841f2-f5ef-4b6a-bc0f-d78b8dbe5347" (UID: "654841f2-f5ef-4b6a-bc0f-d78b8dbe5347"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.280668 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-inventory" (OuterVolumeSpecName: "inventory") pod "654841f2-f5ef-4b6a-bc0f-d78b8dbe5347" (UID: "654841f2-f5ef-4b6a-bc0f-d78b8dbe5347"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.282246 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "654841f2-f5ef-4b6a-bc0f-d78b8dbe5347" (UID: "654841f2-f5ef-4b6a-bc0f-d78b8dbe5347"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.346536 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.346609 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t25sn\" (UniqueName: \"kubernetes.io/projected/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-kube-api-access-t25sn\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.346629 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.346640 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/654841f2-f5ef-4b6a-bc0f-d78b8dbe5347-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.478092 4770 scope.go:117] "RemoveContainer" containerID="dea2502c4d14ad7b8cf534d79544d6e3c6cf8c9762318ff3bb918c1fdd69ee0a" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.509740 4770 scope.go:117] "RemoveContainer" containerID="7bc991c2ef6077dafab86e0be2d1bfbc43aece0619bd6aba7c76104949203229" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.534311 4770 scope.go:117] "RemoveContainer" containerID="010e0458886d5ea64f04f5f881872cd10d6fb70d23fc564fbf462c91b5b28bb6" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.654595 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" event={"ID":"654841f2-f5ef-4b6a-bc0f-d78b8dbe5347","Type":"ContainerDied","Data":"69746dbaed87227163fabcd995ef962fc049c533b3cf38ae3316e2ee22dffcfc"} Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.654635 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69746dbaed87227163fabcd995ef962fc049c533b3cf38ae3316e2ee22dffcfc" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.654672 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-4zjfm" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.737986 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-95slf"] Oct 04 05:16:31 crc kubenswrapper[4770]: E1004 05:16:31.738562 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="654841f2-f5ef-4b6a-bc0f-d78b8dbe5347" containerName="ceph-client-openstack-openstack-cell1" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.738585 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="654841f2-f5ef-4b6a-bc0f-d78b8dbe5347" containerName="ceph-client-openstack-openstack-cell1" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.738792 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="654841f2-f5ef-4b6a-bc0f-d78b8dbe5347" containerName="ceph-client-openstack-openstack-cell1" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.739672 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.741677 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.741948 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.742122 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.742792 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.742954 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.748580 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-95slf"] Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.801931 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.802046 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.802804 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2s92\" (UniqueName: \"kubernetes.io/projected/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-kube-api-access-z2s92\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.803104 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ceph\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.803167 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ssh-key\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.803298 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.803357 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.803380 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-inventory\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.905600 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.905644 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-inventory\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.905757 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2s92\" (UniqueName: \"kubernetes.io/projected/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-kube-api-access-z2s92\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.905867 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ceph\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.905908 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ssh-key\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.905951 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.907747 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.910962 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ssh-key\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.910986 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ceph\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.911095 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-inventory\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.911585 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:31 crc kubenswrapper[4770]: I1004 05:16:31.921120 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2s92\" (UniqueName: \"kubernetes.io/projected/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-kube-api-access-z2s92\") pod \"ovn-openstack-openstack-cell1-95slf\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:32 crc kubenswrapper[4770]: I1004 05:16:32.114216 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:16:32 crc kubenswrapper[4770]: I1004 05:16:32.655106 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-95slf"] Oct 04 05:16:33 crc kubenswrapper[4770]: I1004 05:16:33.686703 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-95slf" event={"ID":"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1","Type":"ContainerStarted","Data":"757548837d898ae0f149867f1dcd4e2248f76c90e9625e7e255c545f40438f84"} Oct 04 05:16:34 crc kubenswrapper[4770]: I1004 05:16:34.694276 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-95slf" event={"ID":"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1","Type":"ContainerStarted","Data":"0b105ce4b6fff779d491cbe578729eccba8f4522f49f987bf682aca48d2c8c25"} Oct 04 05:16:34 crc kubenswrapper[4770]: I1004 05:16:34.717120 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-95slf" podStartSLOduration=2.899170249 podStartE2EDuration="3.717100421s" podCreationTimestamp="2025-10-04 05:16:31 +0000 UTC" firstStartedPulling="2025-10-04 05:16:32.668540493 +0000 UTC m=+8003.960550205" lastFinishedPulling="2025-10-04 05:16:33.486470665 +0000 UTC m=+8004.778480377" observedRunningTime="2025-10-04 05:16:34.708085696 +0000 UTC m=+8006.000095438" watchObservedRunningTime="2025-10-04 05:16:34.717100421 +0000 UTC m=+8006.009110133" Oct 04 05:17:01 crc kubenswrapper[4770]: I1004 05:17:01.795255 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:17:01 crc kubenswrapper[4770]: I1004 05:17:01.795972 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:17:01 crc kubenswrapper[4770]: I1004 05:17:01.796035 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:17:01 crc kubenswrapper[4770]: I1004 05:17:01.796833 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4899d0932e29047d30bb5ffc6e04e928969b5942a1d3a16b4990fd1f4b9ce26f"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:17:01 crc kubenswrapper[4770]: I1004 05:17:01.796878 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://4899d0932e29047d30bb5ffc6e04e928969b5942a1d3a16b4990fd1f4b9ce26f" gracePeriod=600 Oct 04 05:17:01 crc kubenswrapper[4770]: E1004 05:17:01.899556 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bd3be93_1791_4cd1_b3ae_b4032548e93a.slice/crio-4899d0932e29047d30bb5ffc6e04e928969b5942a1d3a16b4990fd1f4b9ce26f.scope\": RecentStats: unable to find data in memory cache]" Oct 04 05:17:01 crc kubenswrapper[4770]: I1004 05:17:01.987947 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="4899d0932e29047d30bb5ffc6e04e928969b5942a1d3a16b4990fd1f4b9ce26f" exitCode=0 Oct 04 05:17:01 crc kubenswrapper[4770]: I1004 05:17:01.988015 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"4899d0932e29047d30bb5ffc6e04e928969b5942a1d3a16b4990fd1f4b9ce26f"} Oct 04 05:17:01 crc kubenswrapper[4770]: I1004 05:17:01.988060 4770 scope.go:117] "RemoveContainer" containerID="2bb5d4af58dd11b0f3518001894de3b24e0a649243937f4cc954fc88950bc253" Oct 04 05:17:03 crc kubenswrapper[4770]: I1004 05:17:03.004757 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93"} Oct 04 05:17:38 crc kubenswrapper[4770]: I1004 05:17:38.347214 4770 generic.go:334] "Generic (PLEG): container finished" podID="535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" containerID="0b105ce4b6fff779d491cbe578729eccba8f4522f49f987bf682aca48d2c8c25" exitCode=0 Oct 04 05:17:38 crc kubenswrapper[4770]: I1004 05:17:38.347311 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-95slf" event={"ID":"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1","Type":"ContainerDied","Data":"0b105ce4b6fff779d491cbe578729eccba8f4522f49f987bf682aca48d2c8c25"} Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.871537 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.977374 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-inventory\") pod \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.977430 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ceph\") pod \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.977540 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovncontroller-config-0\") pod \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.977565 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2s92\" (UniqueName: \"kubernetes.io/projected/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-kube-api-access-z2s92\") pod \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.977680 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovn-combined-ca-bundle\") pod \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.977783 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ssh-key\") pod \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.983901 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ceph" (OuterVolumeSpecName: "ceph") pod "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" (UID: "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.984062 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-kube-api-access-z2s92" (OuterVolumeSpecName: "kube-api-access-z2s92") pod "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" (UID: "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1"). InnerVolumeSpecName "kube-api-access-z2s92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:17:39 crc kubenswrapper[4770]: I1004 05:17:39.985145 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" (UID: "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:17:40 crc kubenswrapper[4770]: E1004 05:17:40.012079 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovncontroller-config-0 podName:535f0f1c-4b20-45bf-b2e2-d16481bfd7c1 nodeName:}" failed. No retries permitted until 2025-10-04 05:17:40.512050735 +0000 UTC m=+8071.804060437 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovncontroller-config-0" (UniqueName: "kubernetes.io/configmap/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovncontroller-config-0") pod "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" (UID: "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1") : error deleting /var/lib/kubelet/pods/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1/volume-subpaths: remove /var/lib/kubelet/pods/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1/volume-subpaths: no such file or directory Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.012542 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-inventory" (OuterVolumeSpecName: "inventory") pod "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" (UID: "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.015111 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" (UID: "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.081191 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.081670 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.081752 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2s92\" (UniqueName: \"kubernetes.io/projected/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-kube-api-access-z2s92\") on node \"crc\" DevicePath \"\"" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.081830 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.081903 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.371437 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-95slf" event={"ID":"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1","Type":"ContainerDied","Data":"757548837d898ae0f149867f1dcd4e2248f76c90e9625e7e255c545f40438f84"} Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.371483 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="757548837d898ae0f149867f1dcd4e2248f76c90e9625e7e255c545f40438f84" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.371534 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-95slf" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.482428 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-stt44"] Oct 04 05:17:40 crc kubenswrapper[4770]: E1004 05:17:40.482928 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" containerName="ovn-openstack-openstack-cell1" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.482946 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" containerName="ovn-openstack-openstack-cell1" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.483151 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" containerName="ovn-openstack-openstack-cell1" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.484084 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.493239 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-stt44"] Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.495984 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.497752 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.591281 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovncontroller-config-0\") pod \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\" (UID: \"535f0f1c-4b20-45bf-b2e2-d16481bfd7c1\") " Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.591819 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1" (UID: "535f0f1c-4b20-45bf-b2e2-d16481bfd7c1"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.592548 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.592722 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.592932 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.593170 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.593363 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.593539 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.593602 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwjhk\" (UniqueName: \"kubernetes.io/projected/c854c1c1-d5eb-471c-9c90-755b2514480c-kube-api-access-bwjhk\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.593815 4770 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/535f0f1c-4b20-45bf-b2e2-d16481bfd7c1-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.695980 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.696061 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.696107 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.696137 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwjhk\" (UniqueName: \"kubernetes.io/projected/c854c1c1-d5eb-471c-9c90-755b2514480c-kube-api-access-bwjhk\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.696171 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.696202 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.696265 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.703026 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.703824 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.706596 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.706835 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.709929 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.716266 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.719076 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwjhk\" (UniqueName: \"kubernetes.io/projected/c854c1c1-d5eb-471c-9c90-755b2514480c-kube-api-access-bwjhk\") pod \"neutron-metadata-openstack-openstack-cell1-stt44\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:40 crc kubenswrapper[4770]: I1004 05:17:40.819204 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:17:41 crc kubenswrapper[4770]: I1004 05:17:41.338410 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-stt44"] Oct 04 05:17:41 crc kubenswrapper[4770]: I1004 05:17:41.385233 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" event={"ID":"c854c1c1-d5eb-471c-9c90-755b2514480c","Type":"ContainerStarted","Data":"072f7eeb6dc4ed68db89a430fe4e3376b28a9b3d14519a647c5290d0151112c6"} Oct 04 05:17:43 crc kubenswrapper[4770]: I1004 05:17:43.407056 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" event={"ID":"c854c1c1-d5eb-471c-9c90-755b2514480c","Type":"ContainerStarted","Data":"123d7a3d9b19a6b85db3383e9369560d5e1cf14aeea8995bff2014eb718839eb"} Oct 04 05:17:43 crc kubenswrapper[4770]: I1004 05:17:43.427596 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" podStartSLOduration=2.152077935 podStartE2EDuration="3.427573484s" podCreationTimestamp="2025-10-04 05:17:40 +0000 UTC" firstStartedPulling="2025-10-04 05:17:41.343437486 +0000 UTC m=+8072.635447198" lastFinishedPulling="2025-10-04 05:17:42.618933035 +0000 UTC m=+8073.910942747" observedRunningTime="2025-10-04 05:17:43.423250731 +0000 UTC m=+8074.715260443" watchObservedRunningTime="2025-10-04 05:17:43.427573484 +0000 UTC m=+8074.719583196" Oct 04 05:18:36 crc kubenswrapper[4770]: I1004 05:18:36.949761 4770 generic.go:334] "Generic (PLEG): container finished" podID="c854c1c1-d5eb-471c-9c90-755b2514480c" containerID="123d7a3d9b19a6b85db3383e9369560d5e1cf14aeea8995bff2014eb718839eb" exitCode=0 Oct 04 05:18:36 crc kubenswrapper[4770]: I1004 05:18:36.949880 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" event={"ID":"c854c1c1-d5eb-471c-9c90-755b2514480c","Type":"ContainerDied","Data":"123d7a3d9b19a6b85db3383e9369560d5e1cf14aeea8995bff2014eb718839eb"} Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.436087 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.478956 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-nova-metadata-neutron-config-0\") pod \"c854c1c1-d5eb-471c-9c90-755b2514480c\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.479697 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwjhk\" (UniqueName: \"kubernetes.io/projected/c854c1c1-d5eb-471c-9c90-755b2514480c-kube-api-access-bwjhk\") pod \"c854c1c1-d5eb-471c-9c90-755b2514480c\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.479802 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-metadata-combined-ca-bundle\") pod \"c854c1c1-d5eb-471c-9c90-755b2514480c\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.479835 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"c854c1c1-d5eb-471c-9c90-755b2514480c\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.479876 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-inventory\") pod \"c854c1c1-d5eb-471c-9c90-755b2514480c\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.479955 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ssh-key\") pod \"c854c1c1-d5eb-471c-9c90-755b2514480c\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.480111 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ceph\") pod \"c854c1c1-d5eb-471c-9c90-755b2514480c\" (UID: \"c854c1c1-d5eb-471c-9c90-755b2514480c\") " Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.487929 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ceph" (OuterVolumeSpecName: "ceph") pod "c854c1c1-d5eb-471c-9c90-755b2514480c" (UID: "c854c1c1-d5eb-471c-9c90-755b2514480c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.500142 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c854c1c1-d5eb-471c-9c90-755b2514480c" (UID: "c854c1c1-d5eb-471c-9c90-755b2514480c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.500291 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c854c1c1-d5eb-471c-9c90-755b2514480c-kube-api-access-bwjhk" (OuterVolumeSpecName: "kube-api-access-bwjhk") pod "c854c1c1-d5eb-471c-9c90-755b2514480c" (UID: "c854c1c1-d5eb-471c-9c90-755b2514480c"). InnerVolumeSpecName "kube-api-access-bwjhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.515139 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "c854c1c1-d5eb-471c-9c90-755b2514480c" (UID: "c854c1c1-d5eb-471c-9c90-755b2514480c"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.517738 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "c854c1c1-d5eb-471c-9c90-755b2514480c" (UID: "c854c1c1-d5eb-471c-9c90-755b2514480c"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.521711 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-inventory" (OuterVolumeSpecName: "inventory") pod "c854c1c1-d5eb-471c-9c90-755b2514480c" (UID: "c854c1c1-d5eb-471c-9c90-755b2514480c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.532929 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c854c1c1-d5eb-471c-9c90-755b2514480c" (UID: "c854c1c1-d5eb-471c-9c90-755b2514480c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.582666 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.582707 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.582717 4770 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.582729 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwjhk\" (UniqueName: \"kubernetes.io/projected/c854c1c1-d5eb-471c-9c90-755b2514480c-kube-api-access-bwjhk\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.582741 4770 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.582755 4770 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.582770 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c854c1c1-d5eb-471c-9c90-755b2514480c-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.972899 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" event={"ID":"c854c1c1-d5eb-471c-9c90-755b2514480c","Type":"ContainerDied","Data":"072f7eeb6dc4ed68db89a430fe4e3376b28a9b3d14519a647c5290d0151112c6"} Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.973262 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="072f7eeb6dc4ed68db89a430fe4e3376b28a9b3d14519a647c5290d0151112c6" Oct 04 05:18:38 crc kubenswrapper[4770]: I1004 05:18:38.972952 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-stt44" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.064050 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-zhbgq"] Oct 04 05:18:39 crc kubenswrapper[4770]: E1004 05:18:39.064664 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c854c1c1-d5eb-471c-9c90-755b2514480c" containerName="neutron-metadata-openstack-openstack-cell1" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.064686 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c854c1c1-d5eb-471c-9c90-755b2514480c" containerName="neutron-metadata-openstack-openstack-cell1" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.064909 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c854c1c1-d5eb-471c-9c90-755b2514480c" containerName="neutron-metadata-openstack-openstack-cell1" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.065730 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.067473 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.068198 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.068277 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.071673 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.072870 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.084704 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-zhbgq"] Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.097589 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.097737 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ceph\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.097797 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ssh-key\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.097961 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.098170 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-inventory\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.098217 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdq66\" (UniqueName: \"kubernetes.io/projected/dab7ef71-5add-4bff-b920-a2b2a974441a-kube-api-access-qdq66\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.200421 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-inventory\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.200482 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdq66\" (UniqueName: \"kubernetes.io/projected/dab7ef71-5add-4bff-b920-a2b2a974441a-kube-api-access-qdq66\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.200621 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.200691 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ceph\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.200729 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ssh-key\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.200778 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.205369 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.205448 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ssh-key\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.205492 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.205931 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-inventory\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.206800 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ceph\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.217619 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdq66\" (UniqueName: \"kubernetes.io/projected/dab7ef71-5add-4bff-b920-a2b2a974441a-kube-api-access-qdq66\") pod \"libvirt-openstack-openstack-cell1-zhbgq\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.385604 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.930377 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-zhbgq"] Oct 04 05:18:39 crc kubenswrapper[4770]: I1004 05:18:39.984413 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" event={"ID":"dab7ef71-5add-4bff-b920-a2b2a974441a","Type":"ContainerStarted","Data":"3aa1b17f475cd3d7e7511142b4ecab9669341b4749e5b0634e738cf604359f15"} Oct 04 05:18:42 crc kubenswrapper[4770]: I1004 05:18:42.008435 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" event={"ID":"dab7ef71-5add-4bff-b920-a2b2a974441a","Type":"ContainerStarted","Data":"8b8434fba9d57ee0f3dd5589b41c6c4840efdb9219bc1fa5e0bd81f38f258aff"} Oct 04 05:18:42 crc kubenswrapper[4770]: I1004 05:18:42.027601 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" podStartSLOduration=1.8982056200000001 podStartE2EDuration="3.027580153s" podCreationTimestamp="2025-10-04 05:18:39 +0000 UTC" firstStartedPulling="2025-10-04 05:18:39.934179459 +0000 UTC m=+8131.226189171" lastFinishedPulling="2025-10-04 05:18:41.063553992 +0000 UTC m=+8132.355563704" observedRunningTime="2025-10-04 05:18:42.023418793 +0000 UTC m=+8133.315428515" watchObservedRunningTime="2025-10-04 05:18:42.027580153 +0000 UTC m=+8133.319589865" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.646789 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nbngw"] Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.650181 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.717823 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbngw"] Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.771857 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxlrn\" (UniqueName: \"kubernetes.io/projected/cea01c9e-3ec8-4558-be9d-8769f70df2ed-kube-api-access-qxlrn\") pod \"redhat-marketplace-nbngw\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.771939 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-catalog-content\") pod \"redhat-marketplace-nbngw\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.772190 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-utilities\") pod \"redhat-marketplace-nbngw\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.875065 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxlrn\" (UniqueName: \"kubernetes.io/projected/cea01c9e-3ec8-4558-be9d-8769f70df2ed-kube-api-access-qxlrn\") pod \"redhat-marketplace-nbngw\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.875148 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-catalog-content\") pod \"redhat-marketplace-nbngw\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.875261 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-utilities\") pod \"redhat-marketplace-nbngw\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.875860 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-utilities\") pod \"redhat-marketplace-nbngw\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.876506 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-catalog-content\") pod \"redhat-marketplace-nbngw\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.894456 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxlrn\" (UniqueName: \"kubernetes.io/projected/cea01c9e-3ec8-4558-be9d-8769f70df2ed-kube-api-access-qxlrn\") pod \"redhat-marketplace-nbngw\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:17 crc kubenswrapper[4770]: I1004 05:19:17.971899 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:18 crc kubenswrapper[4770]: I1004 05:19:18.475209 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbngw"] Oct 04 05:19:19 crc kubenswrapper[4770]: I1004 05:19:19.420475 4770 generic.go:334] "Generic (PLEG): container finished" podID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerID="0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b" exitCode=0 Oct 04 05:19:19 crc kubenswrapper[4770]: I1004 05:19:19.420578 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbngw" event={"ID":"cea01c9e-3ec8-4558-be9d-8769f70df2ed","Type":"ContainerDied","Data":"0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b"} Oct 04 05:19:19 crc kubenswrapper[4770]: I1004 05:19:19.420806 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbngw" event={"ID":"cea01c9e-3ec8-4558-be9d-8769f70df2ed","Type":"ContainerStarted","Data":"ad1019163536bae3446a1dfa3c47ea6fa9fed4bc0edcfb7e278bba9b5229c58d"} Oct 04 05:19:22 crc kubenswrapper[4770]: I1004 05:19:22.456263 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbngw" event={"ID":"cea01c9e-3ec8-4558-be9d-8769f70df2ed","Type":"ContainerStarted","Data":"b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5"} Oct 04 05:19:23 crc kubenswrapper[4770]: I1004 05:19:23.470383 4770 generic.go:334] "Generic (PLEG): container finished" podID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerID="b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5" exitCode=0 Oct 04 05:19:23 crc kubenswrapper[4770]: I1004 05:19:23.470502 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbngw" event={"ID":"cea01c9e-3ec8-4558-be9d-8769f70df2ed","Type":"ContainerDied","Data":"b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5"} Oct 04 05:19:24 crc kubenswrapper[4770]: I1004 05:19:24.485035 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbngw" event={"ID":"cea01c9e-3ec8-4558-be9d-8769f70df2ed","Type":"ContainerStarted","Data":"5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed"} Oct 04 05:19:24 crc kubenswrapper[4770]: I1004 05:19:24.504575 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nbngw" podStartSLOduration=2.859256268 podStartE2EDuration="7.504554912s" podCreationTimestamp="2025-10-04 05:19:17 +0000 UTC" firstStartedPulling="2025-10-04 05:19:19.422465984 +0000 UTC m=+8170.714475696" lastFinishedPulling="2025-10-04 05:19:24.067764628 +0000 UTC m=+8175.359774340" observedRunningTime="2025-10-04 05:19:24.502681612 +0000 UTC m=+8175.794691334" watchObservedRunningTime="2025-10-04 05:19:24.504554912 +0000 UTC m=+8175.796564614" Oct 04 05:19:27 crc kubenswrapper[4770]: I1004 05:19:27.972360 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:27 crc kubenswrapper[4770]: I1004 05:19:27.973032 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:28 crc kubenswrapper[4770]: I1004 05:19:28.024451 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:31 crc kubenswrapper[4770]: I1004 05:19:31.796303 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:19:31 crc kubenswrapper[4770]: I1004 05:19:31.796871 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:19:38 crc kubenswrapper[4770]: I1004 05:19:38.026164 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:38 crc kubenswrapper[4770]: I1004 05:19:38.088757 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbngw"] Oct 04 05:19:38 crc kubenswrapper[4770]: I1004 05:19:38.628231 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nbngw" podUID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerName="registry-server" containerID="cri-o://5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed" gracePeriod=2 Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.167094 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.273044 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxlrn\" (UniqueName: \"kubernetes.io/projected/cea01c9e-3ec8-4558-be9d-8769f70df2ed-kube-api-access-qxlrn\") pod \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.273136 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-utilities\") pod \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.273434 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-catalog-content\") pod \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\" (UID: \"cea01c9e-3ec8-4558-be9d-8769f70df2ed\") " Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.278812 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-utilities" (OuterVolumeSpecName: "utilities") pod "cea01c9e-3ec8-4558-be9d-8769f70df2ed" (UID: "cea01c9e-3ec8-4558-be9d-8769f70df2ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.288264 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cea01c9e-3ec8-4558-be9d-8769f70df2ed-kube-api-access-qxlrn" (OuterVolumeSpecName: "kube-api-access-qxlrn") pod "cea01c9e-3ec8-4558-be9d-8769f70df2ed" (UID: "cea01c9e-3ec8-4558-be9d-8769f70df2ed"). InnerVolumeSpecName "kube-api-access-qxlrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.303825 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cea01c9e-3ec8-4558-be9d-8769f70df2ed" (UID: "cea01c9e-3ec8-4558-be9d-8769f70df2ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.376639 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.376683 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxlrn\" (UniqueName: \"kubernetes.io/projected/cea01c9e-3ec8-4558-be9d-8769f70df2ed-kube-api-access-qxlrn\") on node \"crc\" DevicePath \"\"" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.376694 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea01c9e-3ec8-4558-be9d-8769f70df2ed-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.639129 4770 generic.go:334] "Generic (PLEG): container finished" podID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerID="5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed" exitCode=0 Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.639170 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbngw" event={"ID":"cea01c9e-3ec8-4558-be9d-8769f70df2ed","Type":"ContainerDied","Data":"5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed"} Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.639199 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbngw" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.639215 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbngw" event={"ID":"cea01c9e-3ec8-4558-be9d-8769f70df2ed","Type":"ContainerDied","Data":"ad1019163536bae3446a1dfa3c47ea6fa9fed4bc0edcfb7e278bba9b5229c58d"} Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.639260 4770 scope.go:117] "RemoveContainer" containerID="5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.689402 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbngw"] Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.689714 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbngw"] Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.691829 4770 scope.go:117] "RemoveContainer" containerID="b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.727131 4770 scope.go:117] "RemoveContainer" containerID="0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.765486 4770 scope.go:117] "RemoveContainer" containerID="5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed" Oct 04 05:19:39 crc kubenswrapper[4770]: E1004 05:19:39.766073 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed\": container with ID starting with 5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed not found: ID does not exist" containerID="5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.766122 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed"} err="failed to get container status \"5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed\": rpc error: code = NotFound desc = could not find container \"5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed\": container with ID starting with 5d68d6cf5663fbb7328c913f39f52c548c8425c66fde6bdfdc7a9e85663b94ed not found: ID does not exist" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.766151 4770 scope.go:117] "RemoveContainer" containerID="b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5" Oct 04 05:19:39 crc kubenswrapper[4770]: E1004 05:19:39.766553 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5\": container with ID starting with b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5 not found: ID does not exist" containerID="b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.766595 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5"} err="failed to get container status \"b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5\": rpc error: code = NotFound desc = could not find container \"b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5\": container with ID starting with b304a48c881f57121c51807f38d4a1eb6f701d68e27d469a82c097025e53d0c5 not found: ID does not exist" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.766612 4770 scope.go:117] "RemoveContainer" containerID="0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b" Oct 04 05:19:39 crc kubenswrapper[4770]: E1004 05:19:39.769610 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b\": container with ID starting with 0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b not found: ID does not exist" containerID="0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b" Oct 04 05:19:39 crc kubenswrapper[4770]: I1004 05:19:39.769639 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b"} err="failed to get container status \"0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b\": rpc error: code = NotFound desc = could not find container \"0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b\": container with ID starting with 0a3515a7aad48703f585f651c0a4b58e5f54f482da6ef94933fe2a0eb642ce0b not found: ID does not exist" Oct 04 05:19:41 crc kubenswrapper[4770]: I1004 05:19:41.687167 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" path="/var/lib/kubelet/pods/cea01c9e-3ec8-4558-be9d-8769f70df2ed/volumes" Oct 04 05:20:01 crc kubenswrapper[4770]: I1004 05:20:01.795741 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:20:01 crc kubenswrapper[4770]: I1004 05:20:01.796327 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:20:31 crc kubenswrapper[4770]: I1004 05:20:31.795887 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:20:31 crc kubenswrapper[4770]: I1004 05:20:31.797126 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:20:31 crc kubenswrapper[4770]: I1004 05:20:31.797237 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:20:31 crc kubenswrapper[4770]: I1004 05:20:31.798256 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:20:31 crc kubenswrapper[4770]: I1004 05:20:31.798392 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" gracePeriod=600 Oct 04 05:20:32 crc kubenswrapper[4770]: I1004 05:20:32.178593 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" exitCode=0 Oct 04 05:20:32 crc kubenswrapper[4770]: I1004 05:20:32.178674 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93"} Oct 04 05:20:32 crc kubenswrapper[4770]: I1004 05:20:32.179065 4770 scope.go:117] "RemoveContainer" containerID="4899d0932e29047d30bb5ffc6e04e928969b5942a1d3a16b4990fd1f4b9ce26f" Oct 04 05:20:32 crc kubenswrapper[4770]: E1004 05:20:32.450282 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:20:33 crc kubenswrapper[4770]: I1004 05:20:33.201213 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:20:33 crc kubenswrapper[4770]: E1004 05:20:33.201995 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:20:48 crc kubenswrapper[4770]: I1004 05:20:48.674547 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:20:48 crc kubenswrapper[4770]: E1004 05:20:48.675460 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:21:00 crc kubenswrapper[4770]: I1004 05:21:00.674680 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:21:00 crc kubenswrapper[4770]: E1004 05:21:00.675591 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:21:12 crc kubenswrapper[4770]: I1004 05:21:12.673787 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:21:12 crc kubenswrapper[4770]: E1004 05:21:12.674526 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:21:24 crc kubenswrapper[4770]: I1004 05:21:24.674790 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:21:24 crc kubenswrapper[4770]: E1004 05:21:24.675635 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:21:39 crc kubenswrapper[4770]: I1004 05:21:39.681640 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:21:39 crc kubenswrapper[4770]: E1004 05:21:39.682743 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.641957 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x9tkt"] Oct 04 05:21:47 crc kubenswrapper[4770]: E1004 05:21:47.643145 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerName="extract-utilities" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.643161 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerName="extract-utilities" Oct 04 05:21:47 crc kubenswrapper[4770]: E1004 05:21:47.643172 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerName="extract-content" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.643180 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerName="extract-content" Oct 04 05:21:47 crc kubenswrapper[4770]: E1004 05:21:47.643247 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerName="registry-server" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.643256 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerName="registry-server" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.643496 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="cea01c9e-3ec8-4558-be9d-8769f70df2ed" containerName="registry-server" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.645393 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.655186 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x9tkt"] Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.737285 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-catalog-content\") pod \"community-operators-x9tkt\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.737591 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-utilities\") pod \"community-operators-x9tkt\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.737698 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrptm\" (UniqueName: \"kubernetes.io/projected/6f427363-895e-43a6-bea6-487c8cfae4d7-kube-api-access-vrptm\") pod \"community-operators-x9tkt\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.840293 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-utilities\") pod \"community-operators-x9tkt\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.840357 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrptm\" (UniqueName: \"kubernetes.io/projected/6f427363-895e-43a6-bea6-487c8cfae4d7-kube-api-access-vrptm\") pod \"community-operators-x9tkt\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.840508 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-catalog-content\") pod \"community-operators-x9tkt\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.840907 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-utilities\") pod \"community-operators-x9tkt\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.840942 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-catalog-content\") pod \"community-operators-x9tkt\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.859304 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrptm\" (UniqueName: \"kubernetes.io/projected/6f427363-895e-43a6-bea6-487c8cfae4d7-kube-api-access-vrptm\") pod \"community-operators-x9tkt\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:47 crc kubenswrapper[4770]: I1004 05:21:47.976832 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:21:48 crc kubenswrapper[4770]: I1004 05:21:48.559572 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x9tkt"] Oct 04 05:21:48 crc kubenswrapper[4770]: I1004 05:21:48.981428 4770 generic.go:334] "Generic (PLEG): container finished" podID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerID="d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8" exitCode=0 Oct 04 05:21:48 crc kubenswrapper[4770]: I1004 05:21:48.981548 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x9tkt" event={"ID":"6f427363-895e-43a6-bea6-487c8cfae4d7","Type":"ContainerDied","Data":"d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8"} Oct 04 05:21:48 crc kubenswrapper[4770]: I1004 05:21:48.981801 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x9tkt" event={"ID":"6f427363-895e-43a6-bea6-487c8cfae4d7","Type":"ContainerStarted","Data":"cca557ffb745c298eb81fe994127da3b346b23028e6a83195c43348964d30fab"} Oct 04 05:21:49 crc kubenswrapper[4770]: I1004 05:21:49.992383 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:21:51 crc kubenswrapper[4770]: I1004 05:21:51.673915 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:21:51 crc kubenswrapper[4770]: E1004 05:21:51.674478 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.849476 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kfdrh"] Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.854429 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.863416 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-utilities\") pod \"redhat-operators-kfdrh\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.863462 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-catalog-content\") pod \"redhat-operators-kfdrh\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.863492 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb7p6\" (UniqueName: \"kubernetes.io/projected/af9da074-5402-45db-a4bd-232ad356f21c-kube-api-access-tb7p6\") pod \"redhat-operators-kfdrh\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.884777 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kfdrh"] Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.965597 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-utilities\") pod \"redhat-operators-kfdrh\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.966274 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-catalog-content\") pod \"redhat-operators-kfdrh\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.966222 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-utilities\") pod \"redhat-operators-kfdrh\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.966594 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-catalog-content\") pod \"redhat-operators-kfdrh\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.966879 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb7p6\" (UniqueName: \"kubernetes.io/projected/af9da074-5402-45db-a4bd-232ad356f21c-kube-api-access-tb7p6\") pod \"redhat-operators-kfdrh\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:52 crc kubenswrapper[4770]: I1004 05:21:52.990397 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb7p6\" (UniqueName: \"kubernetes.io/projected/af9da074-5402-45db-a4bd-232ad356f21c-kube-api-access-tb7p6\") pod \"redhat-operators-kfdrh\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:53 crc kubenswrapper[4770]: I1004 05:21:53.192647 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:21:53 crc kubenswrapper[4770]: I1004 05:21:53.719421 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kfdrh"] Oct 04 05:21:53 crc kubenswrapper[4770]: W1004 05:21:53.720038 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf9da074_5402_45db_a4bd_232ad356f21c.slice/crio-7ad56dcc06c747266ee1fb19e80c2f665b49acb457178f2a53ab33411207d076 WatchSource:0}: Error finding container 7ad56dcc06c747266ee1fb19e80c2f665b49acb457178f2a53ab33411207d076: Status 404 returned error can't find the container with id 7ad56dcc06c747266ee1fb19e80c2f665b49acb457178f2a53ab33411207d076 Oct 04 05:21:54 crc kubenswrapper[4770]: I1004 05:21:54.032040 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfdrh" event={"ID":"af9da074-5402-45db-a4bd-232ad356f21c","Type":"ContainerStarted","Data":"7ad56dcc06c747266ee1fb19e80c2f665b49acb457178f2a53ab33411207d076"} Oct 04 05:21:54 crc kubenswrapper[4770]: I1004 05:21:54.034418 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x9tkt" event={"ID":"6f427363-895e-43a6-bea6-487c8cfae4d7","Type":"ContainerStarted","Data":"947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c"} Oct 04 05:21:55 crc kubenswrapper[4770]: I1004 05:21:55.058478 4770 generic.go:334] "Generic (PLEG): container finished" podID="af9da074-5402-45db-a4bd-232ad356f21c" containerID="8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6" exitCode=0 Oct 04 05:21:55 crc kubenswrapper[4770]: I1004 05:21:55.058852 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfdrh" event={"ID":"af9da074-5402-45db-a4bd-232ad356f21c","Type":"ContainerDied","Data":"8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6"} Oct 04 05:22:02 crc kubenswrapper[4770]: I1004 05:22:02.140059 4770 generic.go:334] "Generic (PLEG): container finished" podID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerID="947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c" exitCode=0 Oct 04 05:22:02 crc kubenswrapper[4770]: I1004 05:22:02.140207 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x9tkt" event={"ID":"6f427363-895e-43a6-bea6-487c8cfae4d7","Type":"ContainerDied","Data":"947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c"} Oct 04 05:22:02 crc kubenswrapper[4770]: I1004 05:22:02.145460 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfdrh" event={"ID":"af9da074-5402-45db-a4bd-232ad356f21c","Type":"ContainerStarted","Data":"432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7"} Oct 04 05:22:02 crc kubenswrapper[4770]: I1004 05:22:02.673745 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:22:02 crc kubenswrapper[4770]: E1004 05:22:02.674434 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:22:09 crc kubenswrapper[4770]: I1004 05:22:09.229170 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x9tkt" event={"ID":"6f427363-895e-43a6-bea6-487c8cfae4d7","Type":"ContainerStarted","Data":"2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a"} Oct 04 05:22:09 crc kubenswrapper[4770]: I1004 05:22:09.255642 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x9tkt" podStartSLOduration=4.155717952 podStartE2EDuration="22.255624143s" podCreationTimestamp="2025-10-04 05:21:47 +0000 UTC" firstStartedPulling="2025-10-04 05:21:49.992131304 +0000 UTC m=+8321.284141016" lastFinishedPulling="2025-10-04 05:22:08.092037495 +0000 UTC m=+8339.384047207" observedRunningTime="2025-10-04 05:22:09.246875771 +0000 UTC m=+8340.538885483" watchObservedRunningTime="2025-10-04 05:22:09.255624143 +0000 UTC m=+8340.547633845" Oct 04 05:22:14 crc kubenswrapper[4770]: I1004 05:22:14.674494 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:22:14 crc kubenswrapper[4770]: E1004 05:22:14.675661 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:22:17 crc kubenswrapper[4770]: I1004 05:22:17.977962 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:22:17 crc kubenswrapper[4770]: I1004 05:22:17.978647 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:22:19 crc kubenswrapper[4770]: I1004 05:22:19.024685 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-x9tkt" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="registry-server" probeResult="failure" output=< Oct 04 05:22:19 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:22:19 crc kubenswrapper[4770]: > Oct 04 05:22:26 crc kubenswrapper[4770]: I1004 05:22:26.673581 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:22:26 crc kubenswrapper[4770]: E1004 05:22:26.674616 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:22:29 crc kubenswrapper[4770]: I1004 05:22:29.038452 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-x9tkt" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="registry-server" probeResult="failure" output=< Oct 04 05:22:29 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:22:29 crc kubenswrapper[4770]: > Oct 04 05:22:30 crc kubenswrapper[4770]: I1004 05:22:30.446919 4770 generic.go:334] "Generic (PLEG): container finished" podID="af9da074-5402-45db-a4bd-232ad356f21c" containerID="432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7" exitCode=0 Oct 04 05:22:30 crc kubenswrapper[4770]: I1004 05:22:30.446999 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfdrh" event={"ID":"af9da074-5402-45db-a4bd-232ad356f21c","Type":"ContainerDied","Data":"432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7"} Oct 04 05:22:31 crc kubenswrapper[4770]: I1004 05:22:31.459407 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfdrh" event={"ID":"af9da074-5402-45db-a4bd-232ad356f21c","Type":"ContainerStarted","Data":"233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119"} Oct 04 05:22:31 crc kubenswrapper[4770]: I1004 05:22:31.485469 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kfdrh" podStartSLOduration=3.475503497 podStartE2EDuration="39.485446314s" podCreationTimestamp="2025-10-04 05:21:52 +0000 UTC" firstStartedPulling="2025-10-04 05:21:55.06463758 +0000 UTC m=+8326.356647302" lastFinishedPulling="2025-10-04 05:22:31.074580407 +0000 UTC m=+8362.366590119" observedRunningTime="2025-10-04 05:22:31.475611614 +0000 UTC m=+8362.767621326" watchObservedRunningTime="2025-10-04 05:22:31.485446314 +0000 UTC m=+8362.777456036" Oct 04 05:22:33 crc kubenswrapper[4770]: I1004 05:22:33.193227 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:22:33 crc kubenswrapper[4770]: I1004 05:22:33.194882 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:22:34 crc kubenswrapper[4770]: I1004 05:22:34.251055 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kfdrh" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="registry-server" probeResult="failure" output=< Oct 04 05:22:34 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:22:34 crc kubenswrapper[4770]: > Oct 04 05:22:39 crc kubenswrapper[4770]: I1004 05:22:39.033739 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-x9tkt" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="registry-server" probeResult="failure" output=< Oct 04 05:22:39 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:22:39 crc kubenswrapper[4770]: > Oct 04 05:22:39 crc kubenswrapper[4770]: I1004 05:22:39.681613 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:22:39 crc kubenswrapper[4770]: E1004 05:22:39.681934 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:22:44 crc kubenswrapper[4770]: I1004 05:22:44.240384 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kfdrh" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="registry-server" probeResult="failure" output=< Oct 04 05:22:44 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:22:44 crc kubenswrapper[4770]: > Oct 04 05:22:48 crc kubenswrapper[4770]: I1004 05:22:48.043325 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:22:48 crc kubenswrapper[4770]: I1004 05:22:48.098026 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:22:48 crc kubenswrapper[4770]: I1004 05:22:48.864048 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x9tkt"] Oct 04 05:22:49 crc kubenswrapper[4770]: I1004 05:22:49.631186 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x9tkt" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="registry-server" containerID="cri-o://2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a" gracePeriod=2 Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.175060 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.289463 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-catalog-content\") pod \"6f427363-895e-43a6-bea6-487c8cfae4d7\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.289630 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrptm\" (UniqueName: \"kubernetes.io/projected/6f427363-895e-43a6-bea6-487c8cfae4d7-kube-api-access-vrptm\") pod \"6f427363-895e-43a6-bea6-487c8cfae4d7\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.289657 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-utilities\") pod \"6f427363-895e-43a6-bea6-487c8cfae4d7\" (UID: \"6f427363-895e-43a6-bea6-487c8cfae4d7\") " Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.290718 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-utilities" (OuterVolumeSpecName: "utilities") pod "6f427363-895e-43a6-bea6-487c8cfae4d7" (UID: "6f427363-895e-43a6-bea6-487c8cfae4d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.295937 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f427363-895e-43a6-bea6-487c8cfae4d7-kube-api-access-vrptm" (OuterVolumeSpecName: "kube-api-access-vrptm") pod "6f427363-895e-43a6-bea6-487c8cfae4d7" (UID: "6f427363-895e-43a6-bea6-487c8cfae4d7"). InnerVolumeSpecName "kube-api-access-vrptm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.343962 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f427363-895e-43a6-bea6-487c8cfae4d7" (UID: "6f427363-895e-43a6-bea6-487c8cfae4d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.392542 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrptm\" (UniqueName: \"kubernetes.io/projected/6f427363-895e-43a6-bea6-487c8cfae4d7-kube-api-access-vrptm\") on node \"crc\" DevicePath \"\"" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.392583 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.392592 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f427363-895e-43a6-bea6-487c8cfae4d7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.644676 4770 generic.go:334] "Generic (PLEG): container finished" podID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerID="2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a" exitCode=0 Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.644722 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x9tkt" event={"ID":"6f427363-895e-43a6-bea6-487c8cfae4d7","Type":"ContainerDied","Data":"2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a"} Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.644747 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x9tkt" event={"ID":"6f427363-895e-43a6-bea6-487c8cfae4d7","Type":"ContainerDied","Data":"cca557ffb745c298eb81fe994127da3b346b23028e6a83195c43348964d30fab"} Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.644763 4770 scope.go:117] "RemoveContainer" containerID="2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.644812 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x9tkt" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.675160 4770 scope.go:117] "RemoveContainer" containerID="947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.695751 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x9tkt"] Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.703964 4770 scope.go:117] "RemoveContainer" containerID="d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.704526 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x9tkt"] Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.751854 4770 scope.go:117] "RemoveContainer" containerID="2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a" Oct 04 05:22:50 crc kubenswrapper[4770]: E1004 05:22:50.752462 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a\": container with ID starting with 2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a not found: ID does not exist" containerID="2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.752495 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a"} err="failed to get container status \"2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a\": rpc error: code = NotFound desc = could not find container \"2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a\": container with ID starting with 2e8b4504d373a8fd1760c644bfea70d8795f214d368ce843a7be6b404f51ae1a not found: ID does not exist" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.752516 4770 scope.go:117] "RemoveContainer" containerID="947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c" Oct 04 05:22:50 crc kubenswrapper[4770]: E1004 05:22:50.753002 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c\": container with ID starting with 947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c not found: ID does not exist" containerID="947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.753093 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c"} err="failed to get container status \"947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c\": rpc error: code = NotFound desc = could not find container \"947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c\": container with ID starting with 947269b1d45efe98cea6fb8d552bf2b33ad4d073516b7d844d55c2797ebd805c not found: ID does not exist" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.753153 4770 scope.go:117] "RemoveContainer" containerID="d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8" Oct 04 05:22:50 crc kubenswrapper[4770]: E1004 05:22:50.753696 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8\": container with ID starting with d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8 not found: ID does not exist" containerID="d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8" Oct 04 05:22:50 crc kubenswrapper[4770]: I1004 05:22:50.753898 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8"} err="failed to get container status \"d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8\": rpc error: code = NotFound desc = could not find container \"d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8\": container with ID starting with d495d3c71c81577204ec55a4beb0fc9119464861d8963e4648cda1a1817bdce8 not found: ID does not exist" Oct 04 05:22:51 crc kubenswrapper[4770]: I1004 05:22:51.691521 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" path="/var/lib/kubelet/pods/6f427363-895e-43a6-bea6-487c8cfae4d7/volumes" Oct 04 05:22:53 crc kubenswrapper[4770]: I1004 05:22:53.244616 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:22:53 crc kubenswrapper[4770]: I1004 05:22:53.307864 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:22:54 crc kubenswrapper[4770]: I1004 05:22:54.278708 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kfdrh"] Oct 04 05:22:54 crc kubenswrapper[4770]: I1004 05:22:54.674741 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:22:54 crc kubenswrapper[4770]: E1004 05:22:54.674973 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:22:54 crc kubenswrapper[4770]: I1004 05:22:54.682896 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kfdrh" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="registry-server" containerID="cri-o://233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119" gracePeriod=2 Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.285260 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.310829 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-catalog-content\") pod \"af9da074-5402-45db-a4bd-232ad356f21c\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.315534 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb7p6\" (UniqueName: \"kubernetes.io/projected/af9da074-5402-45db-a4bd-232ad356f21c-kube-api-access-tb7p6\") pod \"af9da074-5402-45db-a4bd-232ad356f21c\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.315846 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-utilities\") pod \"af9da074-5402-45db-a4bd-232ad356f21c\" (UID: \"af9da074-5402-45db-a4bd-232ad356f21c\") " Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.317087 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-utilities" (OuterVolumeSpecName: "utilities") pod "af9da074-5402-45db-a4bd-232ad356f21c" (UID: "af9da074-5402-45db-a4bd-232ad356f21c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.323155 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af9da074-5402-45db-a4bd-232ad356f21c-kube-api-access-tb7p6" (OuterVolumeSpecName: "kube-api-access-tb7p6") pod "af9da074-5402-45db-a4bd-232ad356f21c" (UID: "af9da074-5402-45db-a4bd-232ad356f21c"). InnerVolumeSpecName "kube-api-access-tb7p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.402464 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af9da074-5402-45db-a4bd-232ad356f21c" (UID: "af9da074-5402-45db-a4bd-232ad356f21c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.418298 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.418333 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb7p6\" (UniqueName: \"kubernetes.io/projected/af9da074-5402-45db-a4bd-232ad356f21c-kube-api-access-tb7p6\") on node \"crc\" DevicePath \"\"" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.418347 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9da074-5402-45db-a4bd-232ad356f21c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.697630 4770 generic.go:334] "Generic (PLEG): container finished" podID="af9da074-5402-45db-a4bd-232ad356f21c" containerID="233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119" exitCode=0 Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.697756 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfdrh" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.756986 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfdrh" event={"ID":"af9da074-5402-45db-a4bd-232ad356f21c","Type":"ContainerDied","Data":"233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119"} Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.757163 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kfdrh"] Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.757189 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfdrh" event={"ID":"af9da074-5402-45db-a4bd-232ad356f21c","Type":"ContainerDied","Data":"7ad56dcc06c747266ee1fb19e80c2f665b49acb457178f2a53ab33411207d076"} Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.757217 4770 scope.go:117] "RemoveContainer" containerID="233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.764975 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kfdrh"] Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.781504 4770 scope.go:117] "RemoveContainer" containerID="432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7" Oct 04 05:22:55 crc kubenswrapper[4770]: I1004 05:22:55.814587 4770 scope.go:117] "RemoveContainer" containerID="8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6" Oct 04 05:22:56 crc kubenswrapper[4770]: I1004 05:22:56.006728 4770 scope.go:117] "RemoveContainer" containerID="233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119" Oct 04 05:22:56 crc kubenswrapper[4770]: E1004 05:22:56.007551 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119\": container with ID starting with 233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119 not found: ID does not exist" containerID="233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119" Oct 04 05:22:56 crc kubenswrapper[4770]: I1004 05:22:56.007579 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119"} err="failed to get container status \"233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119\": rpc error: code = NotFound desc = could not find container \"233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119\": container with ID starting with 233a2d57aa92465b76b91d05ee7c273de1fc0b5cb7befd4e351f4f11b87b8119 not found: ID does not exist" Oct 04 05:22:56 crc kubenswrapper[4770]: I1004 05:22:56.007602 4770 scope.go:117] "RemoveContainer" containerID="432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7" Oct 04 05:22:56 crc kubenswrapper[4770]: E1004 05:22:56.008035 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7\": container with ID starting with 432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7 not found: ID does not exist" containerID="432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7" Oct 04 05:22:56 crc kubenswrapper[4770]: I1004 05:22:56.008062 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7"} err="failed to get container status \"432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7\": rpc error: code = NotFound desc = could not find container \"432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7\": container with ID starting with 432c392fddf75e9a7ce5578698c1b13d9e06461de8abeb8db233423da72f00c7 not found: ID does not exist" Oct 04 05:22:56 crc kubenswrapper[4770]: I1004 05:22:56.008077 4770 scope.go:117] "RemoveContainer" containerID="8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6" Oct 04 05:22:56 crc kubenswrapper[4770]: E1004 05:22:56.008450 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6\": container with ID starting with 8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6 not found: ID does not exist" containerID="8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6" Oct 04 05:22:56 crc kubenswrapper[4770]: I1004 05:22:56.008510 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6"} err="failed to get container status \"8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6\": rpc error: code = NotFound desc = could not find container \"8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6\": container with ID starting with 8d098941cb53c22fde7fec5d8a1ecf33ebe0ce0d96f65d45da7e8a5aa052e7d6 not found: ID does not exist" Oct 04 05:22:57 crc kubenswrapper[4770]: I1004 05:22:57.687238 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af9da074-5402-45db-a4bd-232ad356f21c" path="/var/lib/kubelet/pods/af9da074-5402-45db-a4bd-232ad356f21c/volumes" Oct 04 05:23:09 crc kubenswrapper[4770]: I1004 05:23:09.674057 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:23:09 crc kubenswrapper[4770]: E1004 05:23:09.675121 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:23:19 crc kubenswrapper[4770]: I1004 05:23:19.962578 4770 generic.go:334] "Generic (PLEG): container finished" podID="dab7ef71-5add-4bff-b920-a2b2a974441a" containerID="8b8434fba9d57ee0f3dd5589b41c6c4840efdb9219bc1fa5e0bd81f38f258aff" exitCode=0 Oct 04 05:23:19 crc kubenswrapper[4770]: I1004 05:23:19.962658 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" event={"ID":"dab7ef71-5add-4bff-b920-a2b2a974441a","Type":"ContainerDied","Data":"8b8434fba9d57ee0f3dd5589b41c6c4840efdb9219bc1fa5e0bd81f38f258aff"} Oct 04 05:23:20 crc kubenswrapper[4770]: I1004 05:23:20.673705 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:23:20 crc kubenswrapper[4770]: E1004 05:23:20.674052 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.472368 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.599516 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-secret-0\") pod \"dab7ef71-5add-4bff-b920-a2b2a974441a\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.599603 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-combined-ca-bundle\") pod \"dab7ef71-5add-4bff-b920-a2b2a974441a\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.599726 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ssh-key\") pod \"dab7ef71-5add-4bff-b920-a2b2a974441a\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.599842 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdq66\" (UniqueName: \"kubernetes.io/projected/dab7ef71-5add-4bff-b920-a2b2a974441a-kube-api-access-qdq66\") pod \"dab7ef71-5add-4bff-b920-a2b2a974441a\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.599863 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ceph\") pod \"dab7ef71-5add-4bff-b920-a2b2a974441a\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.599881 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-inventory\") pod \"dab7ef71-5add-4bff-b920-a2b2a974441a\" (UID: \"dab7ef71-5add-4bff-b920-a2b2a974441a\") " Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.610336 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "dab7ef71-5add-4bff-b920-a2b2a974441a" (UID: "dab7ef71-5add-4bff-b920-a2b2a974441a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.610458 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ceph" (OuterVolumeSpecName: "ceph") pod "dab7ef71-5add-4bff-b920-a2b2a974441a" (UID: "dab7ef71-5add-4bff-b920-a2b2a974441a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.610463 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab7ef71-5add-4bff-b920-a2b2a974441a-kube-api-access-qdq66" (OuterVolumeSpecName: "kube-api-access-qdq66") pod "dab7ef71-5add-4bff-b920-a2b2a974441a" (UID: "dab7ef71-5add-4bff-b920-a2b2a974441a"). InnerVolumeSpecName "kube-api-access-qdq66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.634203 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dab7ef71-5add-4bff-b920-a2b2a974441a" (UID: "dab7ef71-5add-4bff-b920-a2b2a974441a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.637403 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "dab7ef71-5add-4bff-b920-a2b2a974441a" (UID: "dab7ef71-5add-4bff-b920-a2b2a974441a"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.639729 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-inventory" (OuterVolumeSpecName: "inventory") pod "dab7ef71-5add-4bff-b920-a2b2a974441a" (UID: "dab7ef71-5add-4bff-b920-a2b2a974441a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.701974 4770 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.702021 4770 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.702036 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.702048 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdq66\" (UniqueName: \"kubernetes.io/projected/dab7ef71-5add-4bff-b920-a2b2a974441a-kube-api-access-qdq66\") on node \"crc\" DevicePath \"\"" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.702056 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.702067 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7ef71-5add-4bff-b920-a2b2a974441a-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.988732 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" event={"ID":"dab7ef71-5add-4bff-b920-a2b2a974441a","Type":"ContainerDied","Data":"3aa1b17f475cd3d7e7511142b4ecab9669341b4749e5b0634e738cf604359f15"} Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.990062 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3aa1b17f475cd3d7e7511142b4ecab9669341b4749e5b0634e738cf604359f15" Oct 04 05:23:21 crc kubenswrapper[4770]: I1004 05:23:21.990130 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-zhbgq" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.086658 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-5kbhh"] Oct 04 05:23:22 crc kubenswrapper[4770]: E1004 05:23:22.087160 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="extract-content" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087181 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="extract-content" Oct 04 05:23:22 crc kubenswrapper[4770]: E1004 05:23:22.087216 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="extract-content" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087224 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="extract-content" Oct 04 05:23:22 crc kubenswrapper[4770]: E1004 05:23:22.087260 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="registry-server" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087269 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="registry-server" Oct 04 05:23:22 crc kubenswrapper[4770]: E1004 05:23:22.087287 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab7ef71-5add-4bff-b920-a2b2a974441a" containerName="libvirt-openstack-openstack-cell1" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087296 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab7ef71-5add-4bff-b920-a2b2a974441a" containerName="libvirt-openstack-openstack-cell1" Oct 04 05:23:22 crc kubenswrapper[4770]: E1004 05:23:22.087309 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="extract-utilities" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087318 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="extract-utilities" Oct 04 05:23:22 crc kubenswrapper[4770]: E1004 05:23:22.087338 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="extract-utilities" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087346 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="extract-utilities" Oct 04 05:23:22 crc kubenswrapper[4770]: E1004 05:23:22.087365 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="registry-server" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087373 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="registry-server" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087614 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="af9da074-5402-45db-a4bd-232ad356f21c" containerName="registry-server" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087631 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="dab7ef71-5add-4bff-b920-a2b2a974441a" containerName="libvirt-openstack-openstack-cell1" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.087666 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f427363-895e-43a6-bea6-487c8cfae4d7" containerName="registry-server" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.088531 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.092456 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.092735 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.092889 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.093044 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.093180 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.093295 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.093419 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.101617 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-5kbhh"] Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.214863 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.215076 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.215115 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.215184 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.215234 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.215307 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-868ms\" (UniqueName: \"kubernetes.io/projected/3497238c-e3ec-4169-a636-342eca9f2b8e-kube-api-access-868ms\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.215528 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ceph\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.215817 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.215906 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.215984 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.216109 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-inventory\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.317759 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.317812 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.317833 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.317862 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-inventory\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.317885 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.317967 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.317990 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.318081 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.318160 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.318236 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-868ms\" (UniqueName: \"kubernetes.io/projected/3497238c-e3ec-4169-a636-342eca9f2b8e-kube-api-access-868ms\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.318303 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ceph\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.318881 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.319250 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.323111 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.323571 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.324544 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.324737 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-inventory\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.325197 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.326771 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ceph\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.327663 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.335985 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-868ms\" (UniqueName: \"kubernetes.io/projected/3497238c-e3ec-4169-a636-342eca9f2b8e-kube-api-access-868ms\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.337741 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-5kbhh\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.414607 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.967607 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-5kbhh"] Oct 04 05:23:22 crc kubenswrapper[4770]: I1004 05:23:22.999321 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" event={"ID":"3497238c-e3ec-4169-a636-342eca9f2b8e","Type":"ContainerStarted","Data":"85e32bb10f3e24634d767267ac57afa122bb5e5f0540d000a07a207b68ce752d"} Oct 04 05:23:25 crc kubenswrapper[4770]: I1004 05:23:25.029380 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" event={"ID":"3497238c-e3ec-4169-a636-342eca9f2b8e","Type":"ContainerStarted","Data":"27d3f03b83d1e8abe4cab6eb4973c737d59b7bcefc4244a5070788623a6f175f"} Oct 04 05:23:25 crc kubenswrapper[4770]: I1004 05:23:25.067329 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" podStartSLOduration=1.9578679270000001 podStartE2EDuration="3.067299554s" podCreationTimestamp="2025-10-04 05:23:22 +0000 UTC" firstStartedPulling="2025-10-04 05:23:22.968673232 +0000 UTC m=+8414.260682944" lastFinishedPulling="2025-10-04 05:23:24.078104809 +0000 UTC m=+8415.370114571" observedRunningTime="2025-10-04 05:23:25.050920232 +0000 UTC m=+8416.342929944" watchObservedRunningTime="2025-10-04 05:23:25.067299554 +0000 UTC m=+8416.359309286" Oct 04 05:23:34 crc kubenswrapper[4770]: I1004 05:23:34.673689 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:23:34 crc kubenswrapper[4770]: E1004 05:23:34.675162 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:23:47 crc kubenswrapper[4770]: I1004 05:23:47.675027 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:23:47 crc kubenswrapper[4770]: E1004 05:23:47.676064 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:23:59 crc kubenswrapper[4770]: I1004 05:23:59.674663 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:23:59 crc kubenswrapper[4770]: E1004 05:23:59.675490 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:24:12 crc kubenswrapper[4770]: I1004 05:24:12.675855 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:24:12 crc kubenswrapper[4770]: E1004 05:24:12.676621 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:24:27 crc kubenswrapper[4770]: I1004 05:24:27.673197 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:24:27 crc kubenswrapper[4770]: E1004 05:24:27.674113 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:24:39 crc kubenswrapper[4770]: I1004 05:24:39.687446 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:24:39 crc kubenswrapper[4770]: E1004 05:24:39.688321 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:24:52 crc kubenswrapper[4770]: I1004 05:24:52.674020 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:24:52 crc kubenswrapper[4770]: E1004 05:24:52.675819 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:25:05 crc kubenswrapper[4770]: I1004 05:25:05.674402 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:25:05 crc kubenswrapper[4770]: E1004 05:25:05.675700 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:25:19 crc kubenswrapper[4770]: I1004 05:25:19.687429 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:25:19 crc kubenswrapper[4770]: E1004 05:25:19.688456 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:25:32 crc kubenswrapper[4770]: I1004 05:25:32.673615 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:25:33 crc kubenswrapper[4770]: I1004 05:25:33.336181 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"30fcfb7b289bde5f115704bc9ad874ca96b6a545e50c6cadc0a4fd989cc50fe5"} Oct 04 05:27:08 crc kubenswrapper[4770]: I1004 05:27:08.335374 4770 generic.go:334] "Generic (PLEG): container finished" podID="3497238c-e3ec-4169-a636-342eca9f2b8e" containerID="27d3f03b83d1e8abe4cab6eb4973c737d59b7bcefc4244a5070788623a6f175f" exitCode=0 Oct 04 05:27:08 crc kubenswrapper[4770]: I1004 05:27:08.335461 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" event={"ID":"3497238c-e3ec-4169-a636-342eca9f2b8e","Type":"ContainerDied","Data":"27d3f03b83d1e8abe4cab6eb4973c737d59b7bcefc4244a5070788623a6f175f"} Oct 04 05:27:09 crc kubenswrapper[4770]: I1004 05:27:09.864206 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027349 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-0\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027412 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-inventory\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027504 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-1\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027615 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-combined-ca-bundle\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027658 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-0\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027684 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ssh-key\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027728 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ceph\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027766 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-868ms\" (UniqueName: \"kubernetes.io/projected/3497238c-e3ec-4169-a636-342eca9f2b8e-kube-api-access-868ms\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027834 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-1\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027861 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-0\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.027883 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-1\") pod \"3497238c-e3ec-4169-a636-342eca9f2b8e\" (UID: \"3497238c-e3ec-4169-a636-342eca9f2b8e\") " Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.034888 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3497238c-e3ec-4169-a636-342eca9f2b8e-kube-api-access-868ms" (OuterVolumeSpecName: "kube-api-access-868ms") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "kube-api-access-868ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.035685 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.048319 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ceph" (OuterVolumeSpecName: "ceph") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.059585 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.059857 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.060358 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-inventory" (OuterVolumeSpecName: "inventory") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.063693 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.067749 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.069018 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.074150 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.076059 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "3497238c-e3ec-4169-a636-342eca9f2b8e" (UID: "3497238c-e3ec-4169-a636-342eca9f2b8e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130280 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130321 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130337 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130350 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130361 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130371 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-868ms\" (UniqueName: \"kubernetes.io/projected/3497238c-e3ec-4169-a636-342eca9f2b8e-kube-api-access-868ms\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130379 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130390 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130400 4770 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130409 4770 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.130417 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3497238c-e3ec-4169-a636-342eca9f2b8e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.361904 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" event={"ID":"3497238c-e3ec-4169-a636-342eca9f2b8e","Type":"ContainerDied","Data":"85e32bb10f3e24634d767267ac57afa122bb5e5f0540d000a07a207b68ce752d"} Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.362275 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85e32bb10f3e24634d767267ac57afa122bb5e5f0540d000a07a207b68ce752d" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.361971 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-5kbhh" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.450086 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-bw8b9"] Oct 04 05:27:10 crc kubenswrapper[4770]: E1004 05:27:10.450568 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3497238c-e3ec-4169-a636-342eca9f2b8e" containerName="nova-cell1-openstack-openstack-cell1" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.450590 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3497238c-e3ec-4169-a636-342eca9f2b8e" containerName="nova-cell1-openstack-openstack-cell1" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.450894 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3497238c-e3ec-4169-a636-342eca9f2b8e" containerName="nova-cell1-openstack-openstack-cell1" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.451932 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.453965 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.458650 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.458679 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.458774 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.461865 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.466553 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-bw8b9"] Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.640566 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.640611 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.640651 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.640781 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6528k\" (UniqueName: \"kubernetes.io/projected/a84ee920-d896-418c-afe3-91d4e590e350-kube-api-access-6528k\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.640834 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-inventory\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.640874 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceph\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.640917 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.640932 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ssh-key\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.743231 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.744242 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.744297 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.744625 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6528k\" (UniqueName: \"kubernetes.io/projected/a84ee920-d896-418c-afe3-91d4e590e350-kube-api-access-6528k\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.744684 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-inventory\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.744769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceph\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.744921 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.744946 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ssh-key\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.748520 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-inventory\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.748631 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceph\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.748701 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.749310 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.750521 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.750722 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ssh-key\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.752364 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.762882 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6528k\" (UniqueName: \"kubernetes.io/projected/a84ee920-d896-418c-afe3-91d4e590e350-kube-api-access-6528k\") pod \"telemetry-openstack-openstack-cell1-bw8b9\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:10 crc kubenswrapper[4770]: I1004 05:27:10.776985 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:27:11 crc kubenswrapper[4770]: I1004 05:27:11.330321 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-bw8b9"] Oct 04 05:27:11 crc kubenswrapper[4770]: I1004 05:27:11.340141 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:27:11 crc kubenswrapper[4770]: I1004 05:27:11.378804 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" event={"ID":"a84ee920-d896-418c-afe3-91d4e590e350","Type":"ContainerStarted","Data":"7c0cebc61a52933bb701a8ecfa7e45918ef9d162853884ba2745f631cfcdba5a"} Oct 04 05:27:12 crc kubenswrapper[4770]: I1004 05:27:12.391235 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" event={"ID":"a84ee920-d896-418c-afe3-91d4e590e350","Type":"ContainerStarted","Data":"f4c142ef67ac01680f7f1059883b019e4084a301ad60e8e1ab965123a4224607"} Oct 04 05:27:12 crc kubenswrapper[4770]: I1004 05:27:12.421663 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" podStartSLOduration=1.8321879939999999 podStartE2EDuration="2.421644286s" podCreationTimestamp="2025-10-04 05:27:10 +0000 UTC" firstStartedPulling="2025-10-04 05:27:11.339801259 +0000 UTC m=+8642.631810971" lastFinishedPulling="2025-10-04 05:27:11.929257561 +0000 UTC m=+8643.221267263" observedRunningTime="2025-10-04 05:27:12.410783508 +0000 UTC m=+8643.702793220" watchObservedRunningTime="2025-10-04 05:27:12.421644286 +0000 UTC m=+8643.713653998" Oct 04 05:28:01 crc kubenswrapper[4770]: I1004 05:28:01.796123 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:28:01 crc kubenswrapper[4770]: I1004 05:28:01.796683 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:28:31 crc kubenswrapper[4770]: I1004 05:28:31.795787 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:28:31 crc kubenswrapper[4770]: I1004 05:28:31.796841 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:29:01 crc kubenswrapper[4770]: I1004 05:29:01.795975 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:29:01 crc kubenswrapper[4770]: I1004 05:29:01.796497 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:29:01 crc kubenswrapper[4770]: I1004 05:29:01.796545 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:29:01 crc kubenswrapper[4770]: I1004 05:29:01.798604 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"30fcfb7b289bde5f115704bc9ad874ca96b6a545e50c6cadc0a4fd989cc50fe5"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:29:01 crc kubenswrapper[4770]: I1004 05:29:01.798770 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://30fcfb7b289bde5f115704bc9ad874ca96b6a545e50c6cadc0a4fd989cc50fe5" gracePeriod=600 Oct 04 05:29:02 crc kubenswrapper[4770]: I1004 05:29:02.504183 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="30fcfb7b289bde5f115704bc9ad874ca96b6a545e50c6cadc0a4fd989cc50fe5" exitCode=0 Oct 04 05:29:02 crc kubenswrapper[4770]: I1004 05:29:02.504601 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"30fcfb7b289bde5f115704bc9ad874ca96b6a545e50c6cadc0a4fd989cc50fe5"} Oct 04 05:29:02 crc kubenswrapper[4770]: I1004 05:29:02.504643 4770 scope.go:117] "RemoveContainer" containerID="399293211fbede48845af25c12d0a3c83333d80dfbc815173a051df15b5ebf93" Oct 04 05:29:03 crc kubenswrapper[4770]: I1004 05:29:03.516140 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a"} Oct 04 05:29:56 crc kubenswrapper[4770]: I1004 05:29:56.785648 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xkbd2"] Oct 04 05:29:56 crc kubenswrapper[4770]: I1004 05:29:56.789355 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:56 crc kubenswrapper[4770]: I1004 05:29:56.797385 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkbd2"] Oct 04 05:29:56 crc kubenswrapper[4770]: I1004 05:29:56.939906 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-catalog-content\") pod \"redhat-marketplace-xkbd2\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:56 crc kubenswrapper[4770]: I1004 05:29:56.940093 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-utilities\") pod \"redhat-marketplace-xkbd2\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:56 crc kubenswrapper[4770]: I1004 05:29:56.940204 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95zwv\" (UniqueName: \"kubernetes.io/projected/044bd172-6c27-4d93-af9b-68b47c3cd1df-kube-api-access-95zwv\") pod \"redhat-marketplace-xkbd2\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:57 crc kubenswrapper[4770]: I1004 05:29:57.042431 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-catalog-content\") pod \"redhat-marketplace-xkbd2\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:57 crc kubenswrapper[4770]: I1004 05:29:57.042636 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-utilities\") pod \"redhat-marketplace-xkbd2\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:57 crc kubenswrapper[4770]: I1004 05:29:57.042979 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-catalog-content\") pod \"redhat-marketplace-xkbd2\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:57 crc kubenswrapper[4770]: I1004 05:29:57.043177 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-utilities\") pod \"redhat-marketplace-xkbd2\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:57 crc kubenswrapper[4770]: I1004 05:29:57.043226 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95zwv\" (UniqueName: \"kubernetes.io/projected/044bd172-6c27-4d93-af9b-68b47c3cd1df-kube-api-access-95zwv\") pod \"redhat-marketplace-xkbd2\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:57 crc kubenswrapper[4770]: I1004 05:29:57.064654 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95zwv\" (UniqueName: \"kubernetes.io/projected/044bd172-6c27-4d93-af9b-68b47c3cd1df-kube-api-access-95zwv\") pod \"redhat-marketplace-xkbd2\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:57 crc kubenswrapper[4770]: I1004 05:29:57.118289 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:29:57 crc kubenswrapper[4770]: I1004 05:29:57.602085 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkbd2"] Oct 04 05:29:57 crc kubenswrapper[4770]: W1004 05:29:57.604730 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod044bd172_6c27_4d93_af9b_68b47c3cd1df.slice/crio-f0b39d46bda04c53998a908b6d12f2aeda03d6dd74315e1a3def3898ed7d4e3a WatchSource:0}: Error finding container f0b39d46bda04c53998a908b6d12f2aeda03d6dd74315e1a3def3898ed7d4e3a: Status 404 returned error can't find the container with id f0b39d46bda04c53998a908b6d12f2aeda03d6dd74315e1a3def3898ed7d4e3a Oct 04 05:29:58 crc kubenswrapper[4770]: I1004 05:29:58.079485 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkbd2" event={"ID":"044bd172-6c27-4d93-af9b-68b47c3cd1df","Type":"ContainerStarted","Data":"f0b39d46bda04c53998a908b6d12f2aeda03d6dd74315e1a3def3898ed7d4e3a"} Oct 04 05:29:59 crc kubenswrapper[4770]: I1004 05:29:59.102657 4770 generic.go:334] "Generic (PLEG): container finished" podID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerID="ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0" exitCode=0 Oct 04 05:29:59 crc kubenswrapper[4770]: I1004 05:29:59.102730 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkbd2" event={"ID":"044bd172-6c27-4d93-af9b-68b47c3cd1df","Type":"ContainerDied","Data":"ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0"} Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.116341 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkbd2" event={"ID":"044bd172-6c27-4d93-af9b-68b47c3cd1df","Type":"ContainerStarted","Data":"825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae"} Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.185348 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c"] Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.189145 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.193751 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.194078 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.195971 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c"] Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.323658 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b99654b2-6413-45b9-9c71-e32f2395c35b-config-volume\") pod \"collect-profiles-29325930-7gc2c\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.323777 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5xg2\" (UniqueName: \"kubernetes.io/projected/b99654b2-6413-45b9-9c71-e32f2395c35b-kube-api-access-t5xg2\") pod \"collect-profiles-29325930-7gc2c\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.323829 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b99654b2-6413-45b9-9c71-e32f2395c35b-secret-volume\") pod \"collect-profiles-29325930-7gc2c\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.425784 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b99654b2-6413-45b9-9c71-e32f2395c35b-config-volume\") pod \"collect-profiles-29325930-7gc2c\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.425903 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5xg2\" (UniqueName: \"kubernetes.io/projected/b99654b2-6413-45b9-9c71-e32f2395c35b-kube-api-access-t5xg2\") pod \"collect-profiles-29325930-7gc2c\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.425952 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b99654b2-6413-45b9-9c71-e32f2395c35b-secret-volume\") pod \"collect-profiles-29325930-7gc2c\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.427431 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b99654b2-6413-45b9-9c71-e32f2395c35b-config-volume\") pod \"collect-profiles-29325930-7gc2c\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.432960 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b99654b2-6413-45b9-9c71-e32f2395c35b-secret-volume\") pod \"collect-profiles-29325930-7gc2c\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.453709 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5xg2\" (UniqueName: \"kubernetes.io/projected/b99654b2-6413-45b9-9c71-e32f2395c35b-kube-api-access-t5xg2\") pod \"collect-profiles-29325930-7gc2c\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.522455 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:00 crc kubenswrapper[4770]: I1004 05:30:00.998149 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c"] Oct 04 05:30:01 crc kubenswrapper[4770]: W1004 05:30:01.001397 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb99654b2_6413_45b9_9c71_e32f2395c35b.slice/crio-c00e0b5a48c05f30d47dba1b0d2696f5d47520268403fb9bcf90ebb92bf65ace WatchSource:0}: Error finding container c00e0b5a48c05f30d47dba1b0d2696f5d47520268403fb9bcf90ebb92bf65ace: Status 404 returned error can't find the container with id c00e0b5a48c05f30d47dba1b0d2696f5d47520268403fb9bcf90ebb92bf65ace Oct 04 05:30:01 crc kubenswrapper[4770]: I1004 05:30:01.128183 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" event={"ID":"b99654b2-6413-45b9-9c71-e32f2395c35b","Type":"ContainerStarted","Data":"c00e0b5a48c05f30d47dba1b0d2696f5d47520268403fb9bcf90ebb92bf65ace"} Oct 04 05:30:01 crc kubenswrapper[4770]: I1004 05:30:01.130119 4770 generic.go:334] "Generic (PLEG): container finished" podID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerID="825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae" exitCode=0 Oct 04 05:30:01 crc kubenswrapper[4770]: I1004 05:30:01.130167 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkbd2" event={"ID":"044bd172-6c27-4d93-af9b-68b47c3cd1df","Type":"ContainerDied","Data":"825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae"} Oct 04 05:30:02 crc kubenswrapper[4770]: I1004 05:30:02.141886 4770 generic.go:334] "Generic (PLEG): container finished" podID="b99654b2-6413-45b9-9c71-e32f2395c35b" containerID="813b4d5fddc9fa4d4c58cae454d0f17daa593e4c834fe653204bac7198ff277d" exitCode=0 Oct 04 05:30:02 crc kubenswrapper[4770]: I1004 05:30:02.141939 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" event={"ID":"b99654b2-6413-45b9-9c71-e32f2395c35b","Type":"ContainerDied","Data":"813b4d5fddc9fa4d4c58cae454d0f17daa593e4c834fe653204bac7198ff277d"} Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.157027 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkbd2" event={"ID":"044bd172-6c27-4d93-af9b-68b47c3cd1df","Type":"ContainerStarted","Data":"e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1"} Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.191179 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xkbd2" podStartSLOduration=3.896950736 podStartE2EDuration="7.191159753s" podCreationTimestamp="2025-10-04 05:29:56 +0000 UTC" firstStartedPulling="2025-10-04 05:29:59.106503918 +0000 UTC m=+8810.398513630" lastFinishedPulling="2025-10-04 05:30:02.400712935 +0000 UTC m=+8813.692722647" observedRunningTime="2025-10-04 05:30:03.180207563 +0000 UTC m=+8814.472217285" watchObservedRunningTime="2025-10-04 05:30:03.191159753 +0000 UTC m=+8814.483169475" Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.574934 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.707949 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b99654b2-6413-45b9-9c71-e32f2395c35b-secret-volume\") pod \"b99654b2-6413-45b9-9c71-e32f2395c35b\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.708092 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5xg2\" (UniqueName: \"kubernetes.io/projected/b99654b2-6413-45b9-9c71-e32f2395c35b-kube-api-access-t5xg2\") pod \"b99654b2-6413-45b9-9c71-e32f2395c35b\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.708190 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b99654b2-6413-45b9-9c71-e32f2395c35b-config-volume\") pod \"b99654b2-6413-45b9-9c71-e32f2395c35b\" (UID: \"b99654b2-6413-45b9-9c71-e32f2395c35b\") " Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.708937 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b99654b2-6413-45b9-9c71-e32f2395c35b-config-volume" (OuterVolumeSpecName: "config-volume") pod "b99654b2-6413-45b9-9c71-e32f2395c35b" (UID: "b99654b2-6413-45b9-9c71-e32f2395c35b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.714749 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99654b2-6413-45b9-9c71-e32f2395c35b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b99654b2-6413-45b9-9c71-e32f2395c35b" (UID: "b99654b2-6413-45b9-9c71-e32f2395c35b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.717262 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b99654b2-6413-45b9-9c71-e32f2395c35b-kube-api-access-t5xg2" (OuterVolumeSpecName: "kube-api-access-t5xg2") pod "b99654b2-6413-45b9-9c71-e32f2395c35b" (UID: "b99654b2-6413-45b9-9c71-e32f2395c35b"). InnerVolumeSpecName "kube-api-access-t5xg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.811103 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b99654b2-6413-45b9-9c71-e32f2395c35b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.811143 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5xg2\" (UniqueName: \"kubernetes.io/projected/b99654b2-6413-45b9-9c71-e32f2395c35b-kube-api-access-t5xg2\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:03 crc kubenswrapper[4770]: I1004 05:30:03.811157 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b99654b2-6413-45b9-9c71-e32f2395c35b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:04 crc kubenswrapper[4770]: I1004 05:30:04.171142 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" Oct 04 05:30:04 crc kubenswrapper[4770]: I1004 05:30:04.171116 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-7gc2c" event={"ID":"b99654b2-6413-45b9-9c71-e32f2395c35b","Type":"ContainerDied","Data":"c00e0b5a48c05f30d47dba1b0d2696f5d47520268403fb9bcf90ebb92bf65ace"} Oct 04 05:30:04 crc kubenswrapper[4770]: I1004 05:30:04.171638 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c00e0b5a48c05f30d47dba1b0d2696f5d47520268403fb9bcf90ebb92bf65ace" Oct 04 05:30:04 crc kubenswrapper[4770]: E1004 05:30:04.398229 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb99654b2_6413_45b9_9c71_e32f2395c35b.slice\": RecentStats: unable to find data in memory cache]" Oct 04 05:30:04 crc kubenswrapper[4770]: I1004 05:30:04.650554 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7"] Oct 04 05:30:04 crc kubenswrapper[4770]: I1004 05:30:04.669741 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-gvqx7"] Oct 04 05:30:05 crc kubenswrapper[4770]: I1004 05:30:05.688586 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51eba811-6f6a-4055-9f07-fdcf4ddc9ced" path="/var/lib/kubelet/pods/51eba811-6f6a-4055-9f07-fdcf4ddc9ced/volumes" Oct 04 05:30:07 crc kubenswrapper[4770]: I1004 05:30:07.119768 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:30:07 crc kubenswrapper[4770]: I1004 05:30:07.120177 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:30:07 crc kubenswrapper[4770]: I1004 05:30:07.175175 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:30:07 crc kubenswrapper[4770]: I1004 05:30:07.249788 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:30:07 crc kubenswrapper[4770]: I1004 05:30:07.419421 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkbd2"] Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.223142 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xkbd2" podUID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerName="registry-server" containerID="cri-o://e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1" gracePeriod=2 Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.738450 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.839791 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95zwv\" (UniqueName: \"kubernetes.io/projected/044bd172-6c27-4d93-af9b-68b47c3cd1df-kube-api-access-95zwv\") pod \"044bd172-6c27-4d93-af9b-68b47c3cd1df\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.840172 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-utilities\") pod \"044bd172-6c27-4d93-af9b-68b47c3cd1df\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.840309 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-catalog-content\") pod \"044bd172-6c27-4d93-af9b-68b47c3cd1df\" (UID: \"044bd172-6c27-4d93-af9b-68b47c3cd1df\") " Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.841095 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-utilities" (OuterVolumeSpecName: "utilities") pod "044bd172-6c27-4d93-af9b-68b47c3cd1df" (UID: "044bd172-6c27-4d93-af9b-68b47c3cd1df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.846133 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044bd172-6c27-4d93-af9b-68b47c3cd1df-kube-api-access-95zwv" (OuterVolumeSpecName: "kube-api-access-95zwv") pod "044bd172-6c27-4d93-af9b-68b47c3cd1df" (UID: "044bd172-6c27-4d93-af9b-68b47c3cd1df"). InnerVolumeSpecName "kube-api-access-95zwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.852661 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "044bd172-6c27-4d93-af9b-68b47c3cd1df" (UID: "044bd172-6c27-4d93-af9b-68b47c3cd1df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.942865 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.942922 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044bd172-6c27-4d93-af9b-68b47c3cd1df-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:09 crc kubenswrapper[4770]: I1004 05:30:09.942934 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95zwv\" (UniqueName: \"kubernetes.io/projected/044bd172-6c27-4d93-af9b-68b47c3cd1df-kube-api-access-95zwv\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.234322 4770 generic.go:334] "Generic (PLEG): container finished" podID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerID="e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1" exitCode=0 Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.234379 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkbd2" event={"ID":"044bd172-6c27-4d93-af9b-68b47c3cd1df","Type":"ContainerDied","Data":"e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1"} Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.234412 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkbd2" event={"ID":"044bd172-6c27-4d93-af9b-68b47c3cd1df","Type":"ContainerDied","Data":"f0b39d46bda04c53998a908b6d12f2aeda03d6dd74315e1a3def3898ed7d4e3a"} Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.234415 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkbd2" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.234431 4770 scope.go:117] "RemoveContainer" containerID="e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.268171 4770 scope.go:117] "RemoveContainer" containerID="825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.305280 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkbd2"] Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.314959 4770 scope.go:117] "RemoveContainer" containerID="ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.318220 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkbd2"] Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.355980 4770 scope.go:117] "RemoveContainer" containerID="e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1" Oct 04 05:30:10 crc kubenswrapper[4770]: E1004 05:30:10.356407 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1\": container with ID starting with e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1 not found: ID does not exist" containerID="e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.356447 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1"} err="failed to get container status \"e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1\": rpc error: code = NotFound desc = could not find container \"e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1\": container with ID starting with e02931b9bbbb14192ad0aa4e3956f2f3fe8b0847f0bc28c19ae81126d2676be1 not found: ID does not exist" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.356477 4770 scope.go:117] "RemoveContainer" containerID="825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae" Oct 04 05:30:10 crc kubenswrapper[4770]: E1004 05:30:10.356855 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae\": container with ID starting with 825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae not found: ID does not exist" containerID="825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.356903 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae"} err="failed to get container status \"825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae\": rpc error: code = NotFound desc = could not find container \"825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae\": container with ID starting with 825f989f4dfbfa8d62f6f3de4b58213da32aca903adc554d154377581c00aeae not found: ID does not exist" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.356931 4770 scope.go:117] "RemoveContainer" containerID="ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0" Oct 04 05:30:10 crc kubenswrapper[4770]: E1004 05:30:10.357249 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0\": container with ID starting with ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0 not found: ID does not exist" containerID="ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0" Oct 04 05:30:10 crc kubenswrapper[4770]: I1004 05:30:10.357276 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0"} err="failed to get container status \"ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0\": rpc error: code = NotFound desc = could not find container \"ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0\": container with ID starting with ef59932af1b4b8007e1a178944943139396169765bf6e1cab26f02c38a6484c0 not found: ID does not exist" Oct 04 05:30:11 crc kubenswrapper[4770]: I1004 05:30:11.689934 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="044bd172-6c27-4d93-af9b-68b47c3cd1df" path="/var/lib/kubelet/pods/044bd172-6c27-4d93-af9b-68b47c3cd1df/volumes" Oct 04 05:30:31 crc kubenswrapper[4770]: I1004 05:30:31.957230 4770 scope.go:117] "RemoveContainer" containerID="a10148d1aaaee1d1e1311cba49494469b41b4309448843a0c88d58fd11e18306" Oct 04 05:31:31 crc kubenswrapper[4770]: I1004 05:31:31.795944 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:31:31 crc kubenswrapper[4770]: I1004 05:31:31.796871 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:31:47 crc kubenswrapper[4770]: E1004 05:31:47.296451 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda84ee920_d896_418c_afe3_91d4e590e350.slice/crio-f4c142ef67ac01680f7f1059883b019e4084a301ad60e8e1ab965123a4224607.scope\": RecentStats: unable to find data in memory cache]" Oct 04 05:31:48 crc kubenswrapper[4770]: I1004 05:31:48.202122 4770 generic.go:334] "Generic (PLEG): container finished" podID="a84ee920-d896-418c-afe3-91d4e590e350" containerID="f4c142ef67ac01680f7f1059883b019e4084a301ad60e8e1ab965123a4224607" exitCode=0 Oct 04 05:31:48 crc kubenswrapper[4770]: I1004 05:31:48.202348 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" event={"ID":"a84ee920-d896-418c-afe3-91d4e590e350","Type":"ContainerDied","Data":"f4c142ef67ac01680f7f1059883b019e4084a301ad60e8e1ab965123a4224607"} Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.727288 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.890118 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-1\") pod \"a84ee920-d896-418c-afe3-91d4e590e350\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.890453 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6528k\" (UniqueName: \"kubernetes.io/projected/a84ee920-d896-418c-afe3-91d4e590e350-kube-api-access-6528k\") pod \"a84ee920-d896-418c-afe3-91d4e590e350\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.890550 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceph\") pod \"a84ee920-d896-418c-afe3-91d4e590e350\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.890586 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-inventory\") pod \"a84ee920-d896-418c-afe3-91d4e590e350\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.890749 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-2\") pod \"a84ee920-d896-418c-afe3-91d4e590e350\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.890783 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ssh-key\") pod \"a84ee920-d896-418c-afe3-91d4e590e350\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.890848 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-telemetry-combined-ca-bundle\") pod \"a84ee920-d896-418c-afe3-91d4e590e350\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.890932 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-0\") pod \"a84ee920-d896-418c-afe3-91d4e590e350\" (UID: \"a84ee920-d896-418c-afe3-91d4e590e350\") " Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.896460 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a84ee920-d896-418c-afe3-91d4e590e350" (UID: "a84ee920-d896-418c-afe3-91d4e590e350"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.896584 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceph" (OuterVolumeSpecName: "ceph") pod "a84ee920-d896-418c-afe3-91d4e590e350" (UID: "a84ee920-d896-418c-afe3-91d4e590e350"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.896992 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a84ee920-d896-418c-afe3-91d4e590e350-kube-api-access-6528k" (OuterVolumeSpecName: "kube-api-access-6528k") pod "a84ee920-d896-418c-afe3-91d4e590e350" (UID: "a84ee920-d896-418c-afe3-91d4e590e350"). InnerVolumeSpecName "kube-api-access-6528k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.920540 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-inventory" (OuterVolumeSpecName: "inventory") pod "a84ee920-d896-418c-afe3-91d4e590e350" (UID: "a84ee920-d896-418c-afe3-91d4e590e350"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.921611 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "a84ee920-d896-418c-afe3-91d4e590e350" (UID: "a84ee920-d896-418c-afe3-91d4e590e350"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.922510 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "a84ee920-d896-418c-afe3-91d4e590e350" (UID: "a84ee920-d896-418c-afe3-91d4e590e350"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.923346 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a84ee920-d896-418c-afe3-91d4e590e350" (UID: "a84ee920-d896-418c-afe3-91d4e590e350"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.927554 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "a84ee920-d896-418c-afe3-91d4e590e350" (UID: "a84ee920-d896-418c-afe3-91d4e590e350"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.993490 4770 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.993610 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.993682 4770 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.993737 4770 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.993797 4770 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.993855 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6528k\" (UniqueName: \"kubernetes.io/projected/a84ee920-d896-418c-afe3-91d4e590e350-kube-api-access-6528k\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.993910 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:49 crc kubenswrapper[4770]: I1004 05:31:49.993962 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84ee920-d896-418c-afe3-91d4e590e350-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.223738 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" event={"ID":"a84ee920-d896-418c-afe3-91d4e590e350","Type":"ContainerDied","Data":"7c0cebc61a52933bb701a8ecfa7e45918ef9d162853884ba2745f631cfcdba5a"} Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.224029 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c0cebc61a52933bb701a8ecfa7e45918ef9d162853884ba2745f631cfcdba5a" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.223801 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-bw8b9" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.332877 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-gjmnf"] Oct 04 05:31:50 crc kubenswrapper[4770]: E1004 05:31:50.333514 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerName="extract-utilities" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.333537 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerName="extract-utilities" Oct 04 05:31:50 crc kubenswrapper[4770]: E1004 05:31:50.333555 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerName="extract-content" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.333565 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerName="extract-content" Oct 04 05:31:50 crc kubenswrapper[4770]: E1004 05:31:50.333606 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84ee920-d896-418c-afe3-91d4e590e350" containerName="telemetry-openstack-openstack-cell1" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.333614 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84ee920-d896-418c-afe3-91d4e590e350" containerName="telemetry-openstack-openstack-cell1" Oct 04 05:31:50 crc kubenswrapper[4770]: E1004 05:31:50.333633 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerName="registry-server" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.333641 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerName="registry-server" Oct 04 05:31:50 crc kubenswrapper[4770]: E1004 05:31:50.333660 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99654b2-6413-45b9-9c71-e32f2395c35b" containerName="collect-profiles" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.333669 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99654b2-6413-45b9-9c71-e32f2395c35b" containerName="collect-profiles" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.333901 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="a84ee920-d896-418c-afe3-91d4e590e350" containerName="telemetry-openstack-openstack-cell1" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.333934 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="044bd172-6c27-4d93-af9b-68b47c3cd1df" containerName="registry-server" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.333949 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b99654b2-6413-45b9-9c71-e32f2395c35b" containerName="collect-profiles" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.334918 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.339635 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.339898 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.340124 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.340179 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.340192 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.348126 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-gjmnf"] Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.502823 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.502936 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.502968 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.502989 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.503182 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmcv5\" (UniqueName: \"kubernetes.io/projected/ee2615a0-868c-4c77-ae4c-cfc101b62266-kube-api-access-xmcv5\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.503301 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.605663 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.605705 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.605769 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmcv5\" (UniqueName: \"kubernetes.io/projected/ee2615a0-868c-4c77-ae4c-cfc101b62266-kube-api-access-xmcv5\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.605803 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.605869 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.605975 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.609482 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.609708 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.610212 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.617656 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.624197 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmcv5\" (UniqueName: \"kubernetes.io/projected/ee2615a0-868c-4c77-ae4c-cfc101b62266-kube-api-access-xmcv5\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.624358 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-gjmnf\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:50 crc kubenswrapper[4770]: I1004 05:31:50.673661 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:31:51 crc kubenswrapper[4770]: I1004 05:31:51.264889 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-gjmnf"] Oct 04 05:31:52 crc kubenswrapper[4770]: I1004 05:31:52.246298 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" event={"ID":"ee2615a0-868c-4c77-ae4c-cfc101b62266","Type":"ContainerStarted","Data":"e1fcc1b4185b8850f409af3f16c8cb67b7aa3c1c87452a16d0ed7ddf3d0c01c9"} Oct 04 05:31:55 crc kubenswrapper[4770]: I1004 05:31:55.276374 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" event={"ID":"ee2615a0-868c-4c77-ae4c-cfc101b62266","Type":"ContainerStarted","Data":"6cd68c244a7ea4d57b6aaec08ecda274c00dc302fa6c812bd3a73f990c4ee44b"} Oct 04 05:31:55 crc kubenswrapper[4770]: I1004 05:31:55.301536 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" podStartSLOduration=2.583568563 podStartE2EDuration="5.301514328s" podCreationTimestamp="2025-10-04 05:31:50 +0000 UTC" firstStartedPulling="2025-10-04 05:31:51.265498749 +0000 UTC m=+8922.557508451" lastFinishedPulling="2025-10-04 05:31:53.983444504 +0000 UTC m=+8925.275454216" observedRunningTime="2025-10-04 05:31:55.29440541 +0000 UTC m=+8926.586415122" watchObservedRunningTime="2025-10-04 05:31:55.301514328 +0000 UTC m=+8926.593524040" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.258053 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bf524"] Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.260968 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.277910 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bf524"] Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.362049 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2qdg\" (UniqueName: \"kubernetes.io/projected/39a3f00b-87d7-4580-a778-2bc608aef9ce-kube-api-access-k2qdg\") pod \"community-operators-bf524\" (UID: \"39a3f00b-87d7-4580-a778-2bc608aef9ce\") " pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.362450 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39a3f00b-87d7-4580-a778-2bc608aef9ce-catalog-content\") pod \"community-operators-bf524\" (UID: \"39a3f00b-87d7-4580-a778-2bc608aef9ce\") " pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.362573 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39a3f00b-87d7-4580-a778-2bc608aef9ce-utilities\") pod \"community-operators-bf524\" (UID: \"39a3f00b-87d7-4580-a778-2bc608aef9ce\") " pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.459510 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rhw7s"] Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.467458 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39a3f00b-87d7-4580-a778-2bc608aef9ce-catalog-content\") pod \"community-operators-bf524\" (UID: \"39a3f00b-87d7-4580-a778-2bc608aef9ce\") " pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.467580 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39a3f00b-87d7-4580-a778-2bc608aef9ce-utilities\") pod \"community-operators-bf524\" (UID: \"39a3f00b-87d7-4580-a778-2bc608aef9ce\") " pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.467660 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2qdg\" (UniqueName: \"kubernetes.io/projected/39a3f00b-87d7-4580-a778-2bc608aef9ce-kube-api-access-k2qdg\") pod \"community-operators-bf524\" (UID: \"39a3f00b-87d7-4580-a778-2bc608aef9ce\") " pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.468389 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39a3f00b-87d7-4580-a778-2bc608aef9ce-catalog-content\") pod \"community-operators-bf524\" (UID: \"39a3f00b-87d7-4580-a778-2bc608aef9ce\") " pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.468596 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39a3f00b-87d7-4580-a778-2bc608aef9ce-utilities\") pod \"community-operators-bf524\" (UID: \"39a3f00b-87d7-4580-a778-2bc608aef9ce\") " pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.470995 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.501225 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rhw7s"] Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.504275 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2qdg\" (UniqueName: \"kubernetes.io/projected/39a3f00b-87d7-4580-a778-2bc608aef9ce-kube-api-access-k2qdg\") pod \"community-operators-bf524\" (UID: \"39a3f00b-87d7-4580-a778-2bc608aef9ce\") " pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.569045 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-catalog-content\") pod \"certified-operators-rhw7s\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.569140 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nktjw\" (UniqueName: \"kubernetes.io/projected/aec2dd2c-795f-465a-8cc2-a8cf071787c6-kube-api-access-nktjw\") pod \"certified-operators-rhw7s\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.569208 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-utilities\") pod \"certified-operators-rhw7s\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.589691 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bf524" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.673890 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-catalog-content\") pod \"certified-operators-rhw7s\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.673965 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nktjw\" (UniqueName: \"kubernetes.io/projected/aec2dd2c-795f-465a-8cc2-a8cf071787c6-kube-api-access-nktjw\") pod \"certified-operators-rhw7s\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.674095 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-utilities\") pod \"certified-operators-rhw7s\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.674703 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-utilities\") pod \"certified-operators-rhw7s\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.674968 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-catalog-content\") pod \"certified-operators-rhw7s\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.698712 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nktjw\" (UniqueName: \"kubernetes.io/projected/aec2dd2c-795f-465a-8cc2-a8cf071787c6-kube-api-access-nktjw\") pod \"certified-operators-rhw7s\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:57 crc kubenswrapper[4770]: I1004 05:31:57.801373 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:31:58 crc kubenswrapper[4770]: I1004 05:31:58.324979 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bf524"] Oct 04 05:31:58 crc kubenswrapper[4770]: I1004 05:31:58.403647 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rhw7s"] Oct 04 05:31:58 crc kubenswrapper[4770]: W1004 05:31:58.513096 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaec2dd2c_795f_465a_8cc2_a8cf071787c6.slice/crio-9a07d32089ce554044f5fcb425c3e44d4cc22030b2d8b2364d64e78d1d0f2b6e WatchSource:0}: Error finding container 9a07d32089ce554044f5fcb425c3e44d4cc22030b2d8b2364d64e78d1d0f2b6e: Status 404 returned error can't find the container with id 9a07d32089ce554044f5fcb425c3e44d4cc22030b2d8b2364d64e78d1d0f2b6e Oct 04 05:31:59 crc kubenswrapper[4770]: I1004 05:31:59.317377 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhw7s" event={"ID":"aec2dd2c-795f-465a-8cc2-a8cf071787c6","Type":"ContainerStarted","Data":"9a07d32089ce554044f5fcb425c3e44d4cc22030b2d8b2364d64e78d1d0f2b6e"} Oct 04 05:31:59 crc kubenswrapper[4770]: I1004 05:31:59.318879 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bf524" event={"ID":"39a3f00b-87d7-4580-a778-2bc608aef9ce","Type":"ContainerStarted","Data":"5387f3fd06a42d9fd36f5507a54819e632af040867ecd3a5a21303f3602f2c04"} Oct 04 05:31:59 crc kubenswrapper[4770]: I1004 05:31:59.855652 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9llsw"] Oct 04 05:31:59 crc kubenswrapper[4770]: I1004 05:31:59.858898 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:31:59 crc kubenswrapper[4770]: I1004 05:31:59.874491 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9llsw"] Oct 04 05:31:59 crc kubenswrapper[4770]: I1004 05:31:59.932708 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-catalog-content\") pod \"redhat-operators-9llsw\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:31:59 crc kubenswrapper[4770]: I1004 05:31:59.933064 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-utilities\") pod \"redhat-operators-9llsw\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:31:59 crc kubenswrapper[4770]: I1004 05:31:59.933521 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtw6p\" (UniqueName: \"kubernetes.io/projected/bb238afe-248d-4f68-a026-3ddd72dc9900-kube-api-access-rtw6p\") pod \"redhat-operators-9llsw\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.035703 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-utilities\") pod \"redhat-operators-9llsw\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.036034 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtw6p\" (UniqueName: \"kubernetes.io/projected/bb238afe-248d-4f68-a026-3ddd72dc9900-kube-api-access-rtw6p\") pod \"redhat-operators-9llsw\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.036141 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-catalog-content\") pod \"redhat-operators-9llsw\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.036634 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-catalog-content\") pod \"redhat-operators-9llsw\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.036998 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-utilities\") pod \"redhat-operators-9llsw\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.056934 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtw6p\" (UniqueName: \"kubernetes.io/projected/bb238afe-248d-4f68-a026-3ddd72dc9900-kube-api-access-rtw6p\") pod \"redhat-operators-9llsw\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.202613 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.328804 4770 generic.go:334] "Generic (PLEG): container finished" podID="39a3f00b-87d7-4580-a778-2bc608aef9ce" containerID="5eb44a79e1e7811861b15ef2a864379cafe08d0c3f00d45a7e22e3b56c38d234" exitCode=0 Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.328866 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bf524" event={"ID":"39a3f00b-87d7-4580-a778-2bc608aef9ce","Type":"ContainerDied","Data":"5eb44a79e1e7811861b15ef2a864379cafe08d0c3f00d45a7e22e3b56c38d234"} Oct 04 05:32:00 crc kubenswrapper[4770]: I1004 05:32:00.988037 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9llsw"] Oct 04 05:32:01 crc kubenswrapper[4770]: W1004 05:32:01.220115 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb238afe_248d_4f68_a026_3ddd72dc9900.slice/crio-80f9cd3f110e5e3c038c713dc3652ede7a29af81286f16afb06943555a4f00cd WatchSource:0}: Error finding container 80f9cd3f110e5e3c038c713dc3652ede7a29af81286f16afb06943555a4f00cd: Status 404 returned error can't find the container with id 80f9cd3f110e5e3c038c713dc3652ede7a29af81286f16afb06943555a4f00cd Oct 04 05:32:01 crc kubenswrapper[4770]: I1004 05:32:01.341321 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9llsw" event={"ID":"bb238afe-248d-4f68-a026-3ddd72dc9900","Type":"ContainerStarted","Data":"80f9cd3f110e5e3c038c713dc3652ede7a29af81286f16afb06943555a4f00cd"} Oct 04 05:32:01 crc kubenswrapper[4770]: I1004 05:32:01.347673 4770 generic.go:334] "Generic (PLEG): container finished" podID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerID="b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784" exitCode=0 Oct 04 05:32:01 crc kubenswrapper[4770]: I1004 05:32:01.347821 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhw7s" event={"ID":"aec2dd2c-795f-465a-8cc2-a8cf071787c6","Type":"ContainerDied","Data":"b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784"} Oct 04 05:32:01 crc kubenswrapper[4770]: I1004 05:32:01.795948 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:32:01 crc kubenswrapper[4770]: I1004 05:32:01.796320 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:32:02 crc kubenswrapper[4770]: I1004 05:32:02.362961 4770 generic.go:334] "Generic (PLEG): container finished" podID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerID="2db8c42e4921223c04d13785158340a1e0a1926182a8d0a30702565cd9af82f1" exitCode=0 Oct 04 05:32:02 crc kubenswrapper[4770]: I1004 05:32:02.363024 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9llsw" event={"ID":"bb238afe-248d-4f68-a026-3ddd72dc9900","Type":"ContainerDied","Data":"2db8c42e4921223c04d13785158340a1e0a1926182a8d0a30702565cd9af82f1"} Oct 04 05:32:12 crc kubenswrapper[4770]: I1004 05:32:12.482415 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9llsw" event={"ID":"bb238afe-248d-4f68-a026-3ddd72dc9900","Type":"ContainerStarted","Data":"da05f193a7dac9fc97564fb0281a4b7e7a590bc597113160df45daf3f21fddd6"} Oct 04 05:32:12 crc kubenswrapper[4770]: I1004 05:32:12.484608 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhw7s" event={"ID":"aec2dd2c-795f-465a-8cc2-a8cf071787c6","Type":"ContainerStarted","Data":"8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c"} Oct 04 05:32:12 crc kubenswrapper[4770]: I1004 05:32:12.486837 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bf524" event={"ID":"39a3f00b-87d7-4580-a778-2bc608aef9ce","Type":"ContainerStarted","Data":"b1be0c6a66072c113c92546742e20678ea2fe4ded2692996f65f9e011015a0d0"} Oct 04 05:32:13 crc kubenswrapper[4770]: I1004 05:32:13.502434 4770 generic.go:334] "Generic (PLEG): container finished" podID="39a3f00b-87d7-4580-a778-2bc608aef9ce" containerID="b1be0c6a66072c113c92546742e20678ea2fe4ded2692996f65f9e011015a0d0" exitCode=0 Oct 04 05:32:13 crc kubenswrapper[4770]: I1004 05:32:13.502527 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bf524" event={"ID":"39a3f00b-87d7-4580-a778-2bc608aef9ce","Type":"ContainerDied","Data":"b1be0c6a66072c113c92546742e20678ea2fe4ded2692996f65f9e011015a0d0"} Oct 04 05:32:13 crc kubenswrapper[4770]: I1004 05:32:13.506150 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:32:21 crc kubenswrapper[4770]: I1004 05:32:21.591937 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bf524" event={"ID":"39a3f00b-87d7-4580-a778-2bc608aef9ce","Type":"ContainerStarted","Data":"8cf9e619007e5fdc6aa8f11059fe409c7ed262e40c12f3436d5ac723d51c92bd"} Oct 04 05:32:22 crc kubenswrapper[4770]: I1004 05:32:22.635528 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bf524" podStartSLOduration=7.518729647 podStartE2EDuration="25.635508913s" podCreationTimestamp="2025-10-04 05:31:57 +0000 UTC" firstStartedPulling="2025-10-04 05:32:01.466404569 +0000 UTC m=+8932.758414281" lastFinishedPulling="2025-10-04 05:32:19.583183835 +0000 UTC m=+8950.875193547" observedRunningTime="2025-10-04 05:32:22.630499711 +0000 UTC m=+8953.922509443" watchObservedRunningTime="2025-10-04 05:32:22.635508913 +0000 UTC m=+8953.927518625" Oct 04 05:32:27 crc kubenswrapper[4770]: I1004 05:32:27.590157 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bf524" Oct 04 05:32:27 crc kubenswrapper[4770]: I1004 05:32:27.591219 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bf524" Oct 04 05:32:27 crc kubenswrapper[4770]: I1004 05:32:27.644260 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bf524" Oct 04 05:32:27 crc kubenswrapper[4770]: I1004 05:32:27.701837 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bf524" Oct 04 05:32:28 crc kubenswrapper[4770]: I1004 05:32:28.348306 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bf524"] Oct 04 05:32:28 crc kubenswrapper[4770]: I1004 05:32:28.467574 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-slzl4"] Oct 04 05:32:28 crc kubenswrapper[4770]: I1004 05:32:28.467811 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-slzl4" podUID="8322250a-b542-48c5-a8df-d282483515bb" containerName="registry-server" containerID="cri-o://7d9e936ef71a3e3df3718432cdb01f0d59c56448206548422a1758fde37cf89d" gracePeriod=2 Oct 04 05:32:28 crc kubenswrapper[4770]: E1004 05:32:28.599862 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8322250a_b542_48c5_a8df_d282483515bb.slice/crio-7d9e936ef71a3e3df3718432cdb01f0d59c56448206548422a1758fde37cf89d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8322250a_b542_48c5_a8df_d282483515bb.slice/crio-conmon-7d9e936ef71a3e3df3718432cdb01f0d59c56448206548422a1758fde37cf89d.scope\": RecentStats: unable to find data in memory cache]" Oct 04 05:32:28 crc kubenswrapper[4770]: I1004 05:32:28.711954 4770 generic.go:334] "Generic (PLEG): container finished" podID="8322250a-b542-48c5-a8df-d282483515bb" containerID="7d9e936ef71a3e3df3718432cdb01f0d59c56448206548422a1758fde37cf89d" exitCode=0 Oct 04 05:32:28 crc kubenswrapper[4770]: I1004 05:32:28.712168 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slzl4" event={"ID":"8322250a-b542-48c5-a8df-d282483515bb","Type":"ContainerDied","Data":"7d9e936ef71a3e3df3718432cdb01f0d59c56448206548422a1758fde37cf89d"} Oct 04 05:32:28 crc kubenswrapper[4770]: I1004 05:32:28.716064 4770 generic.go:334] "Generic (PLEG): container finished" podID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerID="8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c" exitCode=0 Oct 04 05:32:28 crc kubenswrapper[4770]: I1004 05:32:28.716397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhw7s" event={"ID":"aec2dd2c-795f-465a-8cc2-a8cf071787c6","Type":"ContainerDied","Data":"8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c"} Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.223798 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slzl4" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.327814 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-utilities\") pod \"8322250a-b542-48c5-a8df-d282483515bb\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.327934 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvkrq\" (UniqueName: \"kubernetes.io/projected/8322250a-b542-48c5-a8df-d282483515bb-kube-api-access-nvkrq\") pod \"8322250a-b542-48c5-a8df-d282483515bb\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.328248 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-catalog-content\") pod \"8322250a-b542-48c5-a8df-d282483515bb\" (UID: \"8322250a-b542-48c5-a8df-d282483515bb\") " Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.328354 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-utilities" (OuterVolumeSpecName: "utilities") pod "8322250a-b542-48c5-a8df-d282483515bb" (UID: "8322250a-b542-48c5-a8df-d282483515bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.329253 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.337640 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8322250a-b542-48c5-a8df-d282483515bb-kube-api-access-nvkrq" (OuterVolumeSpecName: "kube-api-access-nvkrq") pod "8322250a-b542-48c5-a8df-d282483515bb" (UID: "8322250a-b542-48c5-a8df-d282483515bb"). InnerVolumeSpecName "kube-api-access-nvkrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.425301 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8322250a-b542-48c5-a8df-d282483515bb" (UID: "8322250a-b542-48c5-a8df-d282483515bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.431085 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322250a-b542-48c5-a8df-d282483515bb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.431109 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvkrq\" (UniqueName: \"kubernetes.io/projected/8322250a-b542-48c5-a8df-d282483515bb-kube-api-access-nvkrq\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.741533 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slzl4" event={"ID":"8322250a-b542-48c5-a8df-d282483515bb","Type":"ContainerDied","Data":"783fc8ca1452735636382ab4d041a6990bc31fd7dee15efc60a56b5b18ea87b1"} Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.741593 4770 scope.go:117] "RemoveContainer" containerID="7d9e936ef71a3e3df3718432cdb01f0d59c56448206548422a1758fde37cf89d" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.741820 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slzl4" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.748328 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhw7s" event={"ID":"aec2dd2c-795f-465a-8cc2-a8cf071787c6","Type":"ContainerStarted","Data":"d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5"} Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.771524 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-slzl4"] Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.781919 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-slzl4"] Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.790167 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rhw7s" podStartSLOduration=5.885050824 podStartE2EDuration="32.790150115s" podCreationTimestamp="2025-10-04 05:31:57 +0000 UTC" firstStartedPulling="2025-10-04 05:32:02.366247801 +0000 UTC m=+8933.658257513" lastFinishedPulling="2025-10-04 05:32:29.271347092 +0000 UTC m=+8960.563356804" observedRunningTime="2025-10-04 05:32:29.779814881 +0000 UTC m=+8961.071824593" watchObservedRunningTime="2025-10-04 05:32:29.790150115 +0000 UTC m=+8961.082159827" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.801527 4770 scope.go:117] "RemoveContainer" containerID="f741c5e1637129a3f7bf342e9bb3f494d84f14838e4da28cddef19b8c6fe1c1e" Oct 04 05:32:29 crc kubenswrapper[4770]: I1004 05:32:29.871367 4770 scope.go:117] "RemoveContainer" containerID="f29039087d4fd9e9c9f9020bb9dcaf739ab1306a9b6fe949f2761144408bdbd1" Oct 04 05:32:31 crc kubenswrapper[4770]: I1004 05:32:31.685679 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8322250a-b542-48c5-a8df-d282483515bb" path="/var/lib/kubelet/pods/8322250a-b542-48c5-a8df-d282483515bb/volumes" Oct 04 05:32:31 crc kubenswrapper[4770]: I1004 05:32:31.795408 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:32:31 crc kubenswrapper[4770]: I1004 05:32:31.795476 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:32:31 crc kubenswrapper[4770]: I1004 05:32:31.795533 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:32:31 crc kubenswrapper[4770]: I1004 05:32:31.796573 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:32:31 crc kubenswrapper[4770]: I1004 05:32:31.796705 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" gracePeriod=600 Oct 04 05:32:32 crc kubenswrapper[4770]: E1004 05:32:32.435239 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:32:32 crc kubenswrapper[4770]: I1004 05:32:32.783670 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" exitCode=0 Oct 04 05:32:32 crc kubenswrapper[4770]: I1004 05:32:32.783746 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a"} Oct 04 05:32:32 crc kubenswrapper[4770]: I1004 05:32:32.784091 4770 scope.go:117] "RemoveContainer" containerID="30fcfb7b289bde5f115704bc9ad874ca96b6a545e50c6cadc0a4fd989cc50fe5" Oct 04 05:32:32 crc kubenswrapper[4770]: I1004 05:32:32.785652 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:32:32 crc kubenswrapper[4770]: E1004 05:32:32.786588 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:32:35 crc kubenswrapper[4770]: I1004 05:32:35.830184 4770 generic.go:334] "Generic (PLEG): container finished" podID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerID="da05f193a7dac9fc97564fb0281a4b7e7a590bc597113160df45daf3f21fddd6" exitCode=0 Oct 04 05:32:35 crc kubenswrapper[4770]: I1004 05:32:35.830379 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9llsw" event={"ID":"bb238afe-248d-4f68-a026-3ddd72dc9900","Type":"ContainerDied","Data":"da05f193a7dac9fc97564fb0281a4b7e7a590bc597113160df45daf3f21fddd6"} Oct 04 05:32:37 crc kubenswrapper[4770]: I1004 05:32:37.802370 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:32:37 crc kubenswrapper[4770]: I1004 05:32:37.802913 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:32:37 crc kubenswrapper[4770]: I1004 05:32:37.853558 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9llsw" event={"ID":"bb238afe-248d-4f68-a026-3ddd72dc9900","Type":"ContainerStarted","Data":"f7507fc042475d44570ff93b1183dbc42a02952a62040070cbf644751fd7a554"} Oct 04 05:32:37 crc kubenswrapper[4770]: I1004 05:32:37.879143 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9llsw" podStartSLOduration=4.408453349 podStartE2EDuration="38.87912422s" podCreationTimestamp="2025-10-04 05:31:59 +0000 UTC" firstStartedPulling="2025-10-04 05:32:02.366281022 +0000 UTC m=+8933.658290734" lastFinishedPulling="2025-10-04 05:32:36.836951893 +0000 UTC m=+8968.128961605" observedRunningTime="2025-10-04 05:32:37.874747744 +0000 UTC m=+8969.166757456" watchObservedRunningTime="2025-10-04 05:32:37.87912422 +0000 UTC m=+8969.171133932" Oct 04 05:32:38 crc kubenswrapper[4770]: I1004 05:32:38.863872 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rhw7s" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="registry-server" probeResult="failure" output=< Oct 04 05:32:38 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:32:38 crc kubenswrapper[4770]: > Oct 04 05:32:40 crc kubenswrapper[4770]: I1004 05:32:40.203817 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:40 crc kubenswrapper[4770]: I1004 05:32:40.204173 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:32:41 crc kubenswrapper[4770]: I1004 05:32:41.252443 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9llsw" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="registry-server" probeResult="failure" output=< Oct 04 05:32:41 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:32:41 crc kubenswrapper[4770]: > Oct 04 05:32:47 crc kubenswrapper[4770]: I1004 05:32:47.674954 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:32:47 crc kubenswrapper[4770]: E1004 05:32:47.676379 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:32:48 crc kubenswrapper[4770]: I1004 05:32:48.871689 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rhw7s" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="registry-server" probeResult="failure" output=< Oct 04 05:32:48 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:32:48 crc kubenswrapper[4770]: > Oct 04 05:32:51 crc kubenswrapper[4770]: I1004 05:32:51.255835 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9llsw" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="registry-server" probeResult="failure" output=< Oct 04 05:32:51 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:32:51 crc kubenswrapper[4770]: > Oct 04 05:32:57 crc kubenswrapper[4770]: I1004 05:32:57.850877 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:32:57 crc kubenswrapper[4770]: I1004 05:32:57.901844 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:32:58 crc kubenswrapper[4770]: I1004 05:32:58.484262 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rhw7s"] Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.078515 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rhw7s" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="registry-server" containerID="cri-o://d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5" gracePeriod=2 Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.597137 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.653103 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-catalog-content\") pod \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.653329 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nktjw\" (UniqueName: \"kubernetes.io/projected/aec2dd2c-795f-465a-8cc2-a8cf071787c6-kube-api-access-nktjw\") pod \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.653381 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-utilities\") pod \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\" (UID: \"aec2dd2c-795f-465a-8cc2-a8cf071787c6\") " Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.654917 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-utilities" (OuterVolumeSpecName: "utilities") pod "aec2dd2c-795f-465a-8cc2-a8cf071787c6" (UID: "aec2dd2c-795f-465a-8cc2-a8cf071787c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.672779 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aec2dd2c-795f-465a-8cc2-a8cf071787c6-kube-api-access-nktjw" (OuterVolumeSpecName: "kube-api-access-nktjw") pod "aec2dd2c-795f-465a-8cc2-a8cf071787c6" (UID: "aec2dd2c-795f-465a-8cc2-a8cf071787c6"). InnerVolumeSpecName "kube-api-access-nktjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.699968 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aec2dd2c-795f-465a-8cc2-a8cf071787c6" (UID: "aec2dd2c-795f-465a-8cc2-a8cf071787c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.755507 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.755738 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nktjw\" (UniqueName: \"kubernetes.io/projected/aec2dd2c-795f-465a-8cc2-a8cf071787c6-kube-api-access-nktjw\") on node \"crc\" DevicePath \"\"" Oct 04 05:32:59 crc kubenswrapper[4770]: I1004 05:32:59.755754 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec2dd2c-795f-465a-8cc2-a8cf071787c6-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.089118 4770 generic.go:334] "Generic (PLEG): container finished" podID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerID="d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5" exitCode=0 Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.089167 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhw7s" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.089169 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhw7s" event={"ID":"aec2dd2c-795f-465a-8cc2-a8cf071787c6","Type":"ContainerDied","Data":"d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5"} Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.089292 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhw7s" event={"ID":"aec2dd2c-795f-465a-8cc2-a8cf071787c6","Type":"ContainerDied","Data":"9a07d32089ce554044f5fcb425c3e44d4cc22030b2d8b2364d64e78d1d0f2b6e"} Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.089318 4770 scope.go:117] "RemoveContainer" containerID="d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.121523 4770 scope.go:117] "RemoveContainer" containerID="8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.124879 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rhw7s"] Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.134754 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rhw7s"] Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.168269 4770 scope.go:117] "RemoveContainer" containerID="b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.201172 4770 scope.go:117] "RemoveContainer" containerID="d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5" Oct 04 05:33:00 crc kubenswrapper[4770]: E1004 05:33:00.201815 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5\": container with ID starting with d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5 not found: ID does not exist" containerID="d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.201861 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5"} err="failed to get container status \"d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5\": rpc error: code = NotFound desc = could not find container \"d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5\": container with ID starting with d34cf035b1cade84dc7016d6b91c9460926de067139191050ed76e9bf1dc10f5 not found: ID does not exist" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.201894 4770 scope.go:117] "RemoveContainer" containerID="8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c" Oct 04 05:33:00 crc kubenswrapper[4770]: E1004 05:33:00.202324 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c\": container with ID starting with 8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c not found: ID does not exist" containerID="8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.202361 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c"} err="failed to get container status \"8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c\": rpc error: code = NotFound desc = could not find container \"8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c\": container with ID starting with 8a2a4075956fa125c618b7fd66714ef26cae23ae2b2d336ef1444f51cfc4757c not found: ID does not exist" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.202380 4770 scope.go:117] "RemoveContainer" containerID="b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784" Oct 04 05:33:00 crc kubenswrapper[4770]: E1004 05:33:00.202802 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784\": container with ID starting with b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784 not found: ID does not exist" containerID="b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784" Oct 04 05:33:00 crc kubenswrapper[4770]: I1004 05:33:00.202858 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784"} err="failed to get container status \"b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784\": rpc error: code = NotFound desc = could not find container \"b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784\": container with ID starting with b4e24caa0b5ca36f7b2d985b44ad8672cbb85f3b0a8a46fca86fdf1b3841f784 not found: ID does not exist" Oct 04 05:33:01 crc kubenswrapper[4770]: I1004 05:33:01.250776 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9llsw" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="registry-server" probeResult="failure" output=< Oct 04 05:33:01 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:33:01 crc kubenswrapper[4770]: > Oct 04 05:33:01 crc kubenswrapper[4770]: I1004 05:33:01.687521 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" path="/var/lib/kubelet/pods/aec2dd2c-795f-465a-8cc2-a8cf071787c6/volumes" Oct 04 05:33:02 crc kubenswrapper[4770]: I1004 05:33:02.674256 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:33:02 crc kubenswrapper[4770]: E1004 05:33:02.674760 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:33:11 crc kubenswrapper[4770]: I1004 05:33:11.259366 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9llsw" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="registry-server" probeResult="failure" output=< Oct 04 05:33:11 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:33:11 crc kubenswrapper[4770]: > Oct 04 05:33:14 crc kubenswrapper[4770]: I1004 05:33:14.676506 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:33:14 crc kubenswrapper[4770]: E1004 05:33:14.677219 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:33:20 crc kubenswrapper[4770]: I1004 05:33:20.252180 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:33:20 crc kubenswrapper[4770]: I1004 05:33:20.304856 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:33:20 crc kubenswrapper[4770]: I1004 05:33:20.491197 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9llsw"] Oct 04 05:33:21 crc kubenswrapper[4770]: I1004 05:33:21.310090 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9llsw" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="registry-server" containerID="cri-o://f7507fc042475d44570ff93b1183dbc42a02952a62040070cbf644751fd7a554" gracePeriod=2 Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.324341 4770 generic.go:334] "Generic (PLEG): container finished" podID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerID="f7507fc042475d44570ff93b1183dbc42a02952a62040070cbf644751fd7a554" exitCode=0 Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.324549 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9llsw" event={"ID":"bb238afe-248d-4f68-a026-3ddd72dc9900","Type":"ContainerDied","Data":"f7507fc042475d44570ff93b1183dbc42a02952a62040070cbf644751fd7a554"} Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.324883 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9llsw" event={"ID":"bb238afe-248d-4f68-a026-3ddd72dc9900","Type":"ContainerDied","Data":"80f9cd3f110e5e3c038c713dc3652ede7a29af81286f16afb06943555a4f00cd"} Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.324894 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80f9cd3f110e5e3c038c713dc3652ede7a29af81286f16afb06943555a4f00cd" Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.405760 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.561023 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtw6p\" (UniqueName: \"kubernetes.io/projected/bb238afe-248d-4f68-a026-3ddd72dc9900-kube-api-access-rtw6p\") pod \"bb238afe-248d-4f68-a026-3ddd72dc9900\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.561150 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-utilities\") pod \"bb238afe-248d-4f68-a026-3ddd72dc9900\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.561210 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-catalog-content\") pod \"bb238afe-248d-4f68-a026-3ddd72dc9900\" (UID: \"bb238afe-248d-4f68-a026-3ddd72dc9900\") " Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.561913 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-utilities" (OuterVolumeSpecName: "utilities") pod "bb238afe-248d-4f68-a026-3ddd72dc9900" (UID: "bb238afe-248d-4f68-a026-3ddd72dc9900"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.566655 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb238afe-248d-4f68-a026-3ddd72dc9900-kube-api-access-rtw6p" (OuterVolumeSpecName: "kube-api-access-rtw6p") pod "bb238afe-248d-4f68-a026-3ddd72dc9900" (UID: "bb238afe-248d-4f68-a026-3ddd72dc9900"). InnerVolumeSpecName "kube-api-access-rtw6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.638830 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb238afe-248d-4f68-a026-3ddd72dc9900" (UID: "bb238afe-248d-4f68-a026-3ddd72dc9900"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.664333 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtw6p\" (UniqueName: \"kubernetes.io/projected/bb238afe-248d-4f68-a026-3ddd72dc9900-kube-api-access-rtw6p\") on node \"crc\" DevicePath \"\"" Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.664364 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:33:22 crc kubenswrapper[4770]: I1004 05:33:22.664374 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb238afe-248d-4f68-a026-3ddd72dc9900-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:33:23 crc kubenswrapper[4770]: I1004 05:33:23.336207 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9llsw" Oct 04 05:33:23 crc kubenswrapper[4770]: I1004 05:33:23.375687 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9llsw"] Oct 04 05:33:23 crc kubenswrapper[4770]: I1004 05:33:23.386491 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9llsw"] Oct 04 05:33:23 crc kubenswrapper[4770]: I1004 05:33:23.686428 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" path="/var/lib/kubelet/pods/bb238afe-248d-4f68-a026-3ddd72dc9900/volumes" Oct 04 05:33:25 crc kubenswrapper[4770]: I1004 05:33:25.674147 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:33:25 crc kubenswrapper[4770]: E1004 05:33:25.674878 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:33:40 crc kubenswrapper[4770]: I1004 05:33:40.673974 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:33:40 crc kubenswrapper[4770]: E1004 05:33:40.674938 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:33:53 crc kubenswrapper[4770]: I1004 05:33:53.674790 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:33:53 crc kubenswrapper[4770]: E1004 05:33:53.675957 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:34:07 crc kubenswrapper[4770]: I1004 05:34:07.673848 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:34:07 crc kubenswrapper[4770]: E1004 05:34:07.674654 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:34:19 crc kubenswrapper[4770]: I1004 05:34:19.682401 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:34:19 crc kubenswrapper[4770]: E1004 05:34:19.683567 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:34:32 crc kubenswrapper[4770]: I1004 05:34:32.674031 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:34:32 crc kubenswrapper[4770]: E1004 05:34:32.674926 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:34:45 crc kubenswrapper[4770]: I1004 05:34:45.673944 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:34:45 crc kubenswrapper[4770]: E1004 05:34:45.675104 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:34:58 crc kubenswrapper[4770]: I1004 05:34:58.674734 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:34:58 crc kubenswrapper[4770]: E1004 05:34:58.676325 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:35:10 crc kubenswrapper[4770]: I1004 05:35:10.673597 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:35:10 crc kubenswrapper[4770]: E1004 05:35:10.674479 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:35:24 crc kubenswrapper[4770]: I1004 05:35:24.673544 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:35:24 crc kubenswrapper[4770]: E1004 05:35:24.674381 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:35:36 crc kubenswrapper[4770]: I1004 05:35:36.673858 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:35:36 crc kubenswrapper[4770]: E1004 05:35:36.675093 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:35:48 crc kubenswrapper[4770]: I1004 05:35:48.675727 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:35:48 crc kubenswrapper[4770]: E1004 05:35:48.676508 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:36:03 crc kubenswrapper[4770]: I1004 05:36:03.674040 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:36:03 crc kubenswrapper[4770]: E1004 05:36:03.675841 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:36:18 crc kubenswrapper[4770]: I1004 05:36:18.674686 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:36:18 crc kubenswrapper[4770]: E1004 05:36:18.675794 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:36:32 crc kubenswrapper[4770]: I1004 05:36:32.674306 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:36:32 crc kubenswrapper[4770]: E1004 05:36:32.675432 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:36:44 crc kubenswrapper[4770]: I1004 05:36:44.674297 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:36:44 crc kubenswrapper[4770]: E1004 05:36:44.675520 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:36:59 crc kubenswrapper[4770]: I1004 05:36:59.682466 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:36:59 crc kubenswrapper[4770]: E1004 05:36:59.683637 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:37:13 crc kubenswrapper[4770]: I1004 05:37:13.677456 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:37:13 crc kubenswrapper[4770]: E1004 05:37:13.679958 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:37:25 crc kubenswrapper[4770]: I1004 05:37:25.860486 4770 generic.go:334] "Generic (PLEG): container finished" podID="ee2615a0-868c-4c77-ae4c-cfc101b62266" containerID="6cd68c244a7ea4d57b6aaec08ecda274c00dc302fa6c812bd3a73f990c4ee44b" exitCode=0 Oct 04 05:37:25 crc kubenswrapper[4770]: I1004 05:37:25.860652 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" event={"ID":"ee2615a0-868c-4c77-ae4c-cfc101b62266","Type":"ContainerDied","Data":"6cd68c244a7ea4d57b6aaec08ecda274c00dc302fa6c812bd3a73f990c4ee44b"} Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.346491 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.467568 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-combined-ca-bundle\") pod \"ee2615a0-868c-4c77-ae4c-cfc101b62266\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.467965 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-agent-neutron-config-0\") pod \"ee2615a0-868c-4c77-ae4c-cfc101b62266\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.468034 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmcv5\" (UniqueName: \"kubernetes.io/projected/ee2615a0-868c-4c77-ae4c-cfc101b62266-kube-api-access-xmcv5\") pod \"ee2615a0-868c-4c77-ae4c-cfc101b62266\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.468104 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-inventory\") pod \"ee2615a0-868c-4c77-ae4c-cfc101b62266\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.468199 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ssh-key\") pod \"ee2615a0-868c-4c77-ae4c-cfc101b62266\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.468257 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ceph\") pod \"ee2615a0-868c-4c77-ae4c-cfc101b62266\" (UID: \"ee2615a0-868c-4c77-ae4c-cfc101b62266\") " Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.474958 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee2615a0-868c-4c77-ae4c-cfc101b62266-kube-api-access-xmcv5" (OuterVolumeSpecName: "kube-api-access-xmcv5") pod "ee2615a0-868c-4c77-ae4c-cfc101b62266" (UID: "ee2615a0-868c-4c77-ae4c-cfc101b62266"). InnerVolumeSpecName "kube-api-access-xmcv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.475423 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ceph" (OuterVolumeSpecName: "ceph") pod "ee2615a0-868c-4c77-ae4c-cfc101b62266" (UID: "ee2615a0-868c-4c77-ae4c-cfc101b62266"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.475498 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "ee2615a0-868c-4c77-ae4c-cfc101b62266" (UID: "ee2615a0-868c-4c77-ae4c-cfc101b62266"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.497875 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ee2615a0-868c-4c77-ae4c-cfc101b62266" (UID: "ee2615a0-868c-4c77-ae4c-cfc101b62266"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.500741 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-inventory" (OuterVolumeSpecName: "inventory") pod "ee2615a0-868c-4c77-ae4c-cfc101b62266" (UID: "ee2615a0-868c-4c77-ae4c-cfc101b62266"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.515232 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "ee2615a0-868c-4c77-ae4c-cfc101b62266" (UID: "ee2615a0-868c-4c77-ae4c-cfc101b62266"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.575060 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.575096 4770 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.575112 4770 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.575128 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmcv5\" (UniqueName: \"kubernetes.io/projected/ee2615a0-868c-4c77-ae4c-cfc101b62266-kube-api-access-xmcv5\") on node \"crc\" DevicePath \"\"" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.575140 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.575151 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee2615a0-868c-4c77-ae4c-cfc101b62266-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.674193 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:37:27 crc kubenswrapper[4770]: E1004 05:37:27.674685 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.883103 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" event={"ID":"ee2615a0-868c-4c77-ae4c-cfc101b62266","Type":"ContainerDied","Data":"e1fcc1b4185b8850f409af3f16c8cb67b7aa3c1c87452a16d0ed7ddf3d0c01c9"} Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.883161 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1fcc1b4185b8850f409af3f16c8cb67b7aa3c1c87452a16d0ed7ddf3d0c01c9" Oct 04 05:37:27 crc kubenswrapper[4770]: I1004 05:37:27.883238 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-gjmnf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.007751 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-68ssf"] Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008266 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="extract-content" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008293 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="extract-content" Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008320 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="registry-server" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008328 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="registry-server" Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008345 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="extract-utilities" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008354 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="extract-utilities" Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008401 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322250a-b542-48c5-a8df-d282483515bb" containerName="registry-server" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008409 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322250a-b542-48c5-a8df-d282483515bb" containerName="registry-server" Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008425 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="extract-utilities" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008432 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="extract-utilities" Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008441 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="extract-content" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008448 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="extract-content" Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008464 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="registry-server" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008472 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="registry-server" Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008486 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2615a0-868c-4c77-ae4c-cfc101b62266" containerName="neutron-sriov-openstack-openstack-cell1" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008493 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2615a0-868c-4c77-ae4c-cfc101b62266" containerName="neutron-sriov-openstack-openstack-cell1" Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008505 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322250a-b542-48c5-a8df-d282483515bb" containerName="extract-content" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008511 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322250a-b542-48c5-a8df-d282483515bb" containerName="extract-content" Oct 04 05:37:28 crc kubenswrapper[4770]: E1004 05:37:28.008520 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322250a-b542-48c5-a8df-d282483515bb" containerName="extract-utilities" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008527 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322250a-b542-48c5-a8df-d282483515bb" containerName="extract-utilities" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008793 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="aec2dd2c-795f-465a-8cc2-a8cf071787c6" containerName="registry-server" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008812 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb238afe-248d-4f68-a026-3ddd72dc9900" containerName="registry-server" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008826 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2615a0-868c-4c77-ae4c-cfc101b62266" containerName="neutron-sriov-openstack-openstack-cell1" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.008845 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8322250a-b542-48c5-a8df-d282483515bb" containerName="registry-server" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.009726 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.013478 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.013706 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.013844 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.015211 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.018352 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.021245 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-68ssf"] Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.188505 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.188627 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.188748 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.188949 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.189145 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbzp2\" (UniqueName: \"kubernetes.io/projected/6707d041-c3e2-4445-b381-ce39c1e572a8-kube-api-access-mbzp2\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.189209 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.291791 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.291871 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.291920 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.291967 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.292030 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbzp2\" (UniqueName: \"kubernetes.io/projected/6707d041-c3e2-4445-b381-ce39c1e572a8-kube-api-access-mbzp2\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.292069 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.295865 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.296409 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.296544 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.296626 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.297052 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.311880 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbzp2\" (UniqueName: \"kubernetes.io/projected/6707d041-c3e2-4445-b381-ce39c1e572a8-kube-api-access-mbzp2\") pod \"neutron-dhcp-openstack-openstack-cell1-68ssf\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.333468 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:37:28 crc kubenswrapper[4770]: W1004 05:37:28.913101 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6707d041_c3e2_4445_b381_ce39c1e572a8.slice/crio-85adb7f2d11a33f98675bd926a8841b5d5750aa04cfe9ccb32eeefa9406aff1d WatchSource:0}: Error finding container 85adb7f2d11a33f98675bd926a8841b5d5750aa04cfe9ccb32eeefa9406aff1d: Status 404 returned error can't find the container with id 85adb7f2d11a33f98675bd926a8841b5d5750aa04cfe9ccb32eeefa9406aff1d Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.913358 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-68ssf"] Oct 04 05:37:28 crc kubenswrapper[4770]: I1004 05:37:28.915660 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:37:29 crc kubenswrapper[4770]: I1004 05:37:29.903893 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" event={"ID":"6707d041-c3e2-4445-b381-ce39c1e572a8","Type":"ContainerStarted","Data":"85adb7f2d11a33f98675bd926a8841b5d5750aa04cfe9ccb32eeefa9406aff1d"} Oct 04 05:37:30 crc kubenswrapper[4770]: I1004 05:37:30.919710 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" event={"ID":"6707d041-c3e2-4445-b381-ce39c1e572a8","Type":"ContainerStarted","Data":"b38f32eb7dcbd1d9c3e733da028380e28a5891763e81fc9aee49e3f7ce31f59e"} Oct 04 05:37:30 crc kubenswrapper[4770]: I1004 05:37:30.936313 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" podStartSLOduration=3.46625482 podStartE2EDuration="3.936291925s" podCreationTimestamp="2025-10-04 05:37:27 +0000 UTC" firstStartedPulling="2025-10-04 05:37:28.915417408 +0000 UTC m=+9260.207427120" lastFinishedPulling="2025-10-04 05:37:29.385454513 +0000 UTC m=+9260.677464225" observedRunningTime="2025-10-04 05:37:30.933456158 +0000 UTC m=+9262.225465880" watchObservedRunningTime="2025-10-04 05:37:30.936291925 +0000 UTC m=+9262.228301667" Oct 04 05:37:38 crc kubenswrapper[4770]: I1004 05:37:38.674819 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:37:39 crc kubenswrapper[4770]: I1004 05:37:39.012370 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"3ecefd076d2acf3bcf8b76829f054fa851e70adb6ef01cc6778eaf1cc5cb4ce6"} Oct 04 05:38:32 crc kubenswrapper[4770]: I1004 05:38:32.254952 4770 scope.go:117] "RemoveContainer" containerID="2db8c42e4921223c04d13785158340a1e0a1926182a8d0a30702565cd9af82f1" Oct 04 05:38:32 crc kubenswrapper[4770]: I1004 05:38:32.290663 4770 scope.go:117] "RemoveContainer" containerID="da05f193a7dac9fc97564fb0281a4b7e7a590bc597113160df45daf3f21fddd6" Oct 04 05:39:32 crc kubenswrapper[4770]: I1004 05:39:32.359282 4770 scope.go:117] "RemoveContainer" containerID="f7507fc042475d44570ff93b1183dbc42a02952a62040070cbf644751fd7a554" Oct 04 05:40:01 crc kubenswrapper[4770]: I1004 05:40:01.795345 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:40:01 crc kubenswrapper[4770]: I1004 05:40:01.795868 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.045671 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v7ftf"] Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.048991 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.061609 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7ftf"] Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.129208 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-catalog-content\") pod \"redhat-marketplace-v7ftf\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.129283 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-utilities\") pod \"redhat-marketplace-v7ftf\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.129386 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkz89\" (UniqueName: \"kubernetes.io/projected/8631a507-43e1-4923-a057-54f930b3fdab-kube-api-access-wkz89\") pod \"redhat-marketplace-v7ftf\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.231641 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkz89\" (UniqueName: \"kubernetes.io/projected/8631a507-43e1-4923-a057-54f930b3fdab-kube-api-access-wkz89\") pod \"redhat-marketplace-v7ftf\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.231811 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-catalog-content\") pod \"redhat-marketplace-v7ftf\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.231856 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-utilities\") pod \"redhat-marketplace-v7ftf\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.232408 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-catalog-content\") pod \"redhat-marketplace-v7ftf\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.232464 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-utilities\") pod \"redhat-marketplace-v7ftf\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.254375 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkz89\" (UniqueName: \"kubernetes.io/projected/8631a507-43e1-4923-a057-54f930b3fdab-kube-api-access-wkz89\") pod \"redhat-marketplace-v7ftf\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.372583 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:23 crc kubenswrapper[4770]: I1004 05:40:23.837416 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7ftf"] Oct 04 05:40:24 crc kubenswrapper[4770]: I1004 05:40:24.686935 4770 generic.go:334] "Generic (PLEG): container finished" podID="8631a507-43e1-4923-a057-54f930b3fdab" containerID="3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e" exitCode=0 Oct 04 05:40:24 crc kubenswrapper[4770]: I1004 05:40:24.687067 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7ftf" event={"ID":"8631a507-43e1-4923-a057-54f930b3fdab","Type":"ContainerDied","Data":"3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e"} Oct 04 05:40:24 crc kubenswrapper[4770]: I1004 05:40:24.687323 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7ftf" event={"ID":"8631a507-43e1-4923-a057-54f930b3fdab","Type":"ContainerStarted","Data":"52521490fc42686249b1e98237ff169d6a3d2e8f039b46f41d46f06b7cc2b8eb"} Oct 04 05:40:26 crc kubenswrapper[4770]: I1004 05:40:26.709671 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7ftf" event={"ID":"8631a507-43e1-4923-a057-54f930b3fdab","Type":"ContainerStarted","Data":"38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd"} Oct 04 05:40:27 crc kubenswrapper[4770]: I1004 05:40:27.723431 4770 generic.go:334] "Generic (PLEG): container finished" podID="8631a507-43e1-4923-a057-54f930b3fdab" containerID="38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd" exitCode=0 Oct 04 05:40:27 crc kubenswrapper[4770]: I1004 05:40:27.723614 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7ftf" event={"ID":"8631a507-43e1-4923-a057-54f930b3fdab","Type":"ContainerDied","Data":"38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd"} Oct 04 05:40:29 crc kubenswrapper[4770]: I1004 05:40:29.744403 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7ftf" event={"ID":"8631a507-43e1-4923-a057-54f930b3fdab","Type":"ContainerStarted","Data":"9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a"} Oct 04 05:40:29 crc kubenswrapper[4770]: I1004 05:40:29.761088 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v7ftf" podStartSLOduration=3.053717878 podStartE2EDuration="6.76107018s" podCreationTimestamp="2025-10-04 05:40:23 +0000 UTC" firstStartedPulling="2025-10-04 05:40:24.71337528 +0000 UTC m=+9436.005384992" lastFinishedPulling="2025-10-04 05:40:28.420727582 +0000 UTC m=+9439.712737294" observedRunningTime="2025-10-04 05:40:29.760383172 +0000 UTC m=+9441.052392884" watchObservedRunningTime="2025-10-04 05:40:29.76107018 +0000 UTC m=+9441.053079902" Oct 04 05:40:31 crc kubenswrapper[4770]: I1004 05:40:31.796290 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:40:31 crc kubenswrapper[4770]: I1004 05:40:31.798063 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:40:33 crc kubenswrapper[4770]: I1004 05:40:33.373061 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:33 crc kubenswrapper[4770]: I1004 05:40:33.374815 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:33 crc kubenswrapper[4770]: I1004 05:40:33.424881 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:34 crc kubenswrapper[4770]: I1004 05:40:34.840177 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:34 crc kubenswrapper[4770]: I1004 05:40:34.889541 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7ftf"] Oct 04 05:40:36 crc kubenswrapper[4770]: I1004 05:40:36.819193 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v7ftf" podUID="8631a507-43e1-4923-a057-54f930b3fdab" containerName="registry-server" containerID="cri-o://9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a" gracePeriod=2 Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.308364 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.339282 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkz89\" (UniqueName: \"kubernetes.io/projected/8631a507-43e1-4923-a057-54f930b3fdab-kube-api-access-wkz89\") pod \"8631a507-43e1-4923-a057-54f930b3fdab\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.339446 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-catalog-content\") pod \"8631a507-43e1-4923-a057-54f930b3fdab\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.339477 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-utilities\") pod \"8631a507-43e1-4923-a057-54f930b3fdab\" (UID: \"8631a507-43e1-4923-a057-54f930b3fdab\") " Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.340680 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-utilities" (OuterVolumeSpecName: "utilities") pod "8631a507-43e1-4923-a057-54f930b3fdab" (UID: "8631a507-43e1-4923-a057-54f930b3fdab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.351303 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8631a507-43e1-4923-a057-54f930b3fdab-kube-api-access-wkz89" (OuterVolumeSpecName: "kube-api-access-wkz89") pod "8631a507-43e1-4923-a057-54f930b3fdab" (UID: "8631a507-43e1-4923-a057-54f930b3fdab"). InnerVolumeSpecName "kube-api-access-wkz89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.353965 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8631a507-43e1-4923-a057-54f930b3fdab" (UID: "8631a507-43e1-4923-a057-54f930b3fdab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.442027 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.442065 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8631a507-43e1-4923-a057-54f930b3fdab-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.442078 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkz89\" (UniqueName: \"kubernetes.io/projected/8631a507-43e1-4923-a057-54f930b3fdab-kube-api-access-wkz89\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.829796 4770 generic.go:334] "Generic (PLEG): container finished" podID="8631a507-43e1-4923-a057-54f930b3fdab" containerID="9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a" exitCode=0 Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.829862 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7ftf" event={"ID":"8631a507-43e1-4923-a057-54f930b3fdab","Type":"ContainerDied","Data":"9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a"} Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.829871 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7ftf" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.830270 4770 scope.go:117] "RemoveContainer" containerID="9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.830258 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7ftf" event={"ID":"8631a507-43e1-4923-a057-54f930b3fdab","Type":"ContainerDied","Data":"52521490fc42686249b1e98237ff169d6a3d2e8f039b46f41d46f06b7cc2b8eb"} Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.855982 4770 scope.go:117] "RemoveContainer" containerID="38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.882128 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7ftf"] Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.882236 4770 scope.go:117] "RemoveContainer" containerID="3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.892326 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7ftf"] Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.926407 4770 scope.go:117] "RemoveContainer" containerID="9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a" Oct 04 05:40:37 crc kubenswrapper[4770]: E1004 05:40:37.926784 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a\": container with ID starting with 9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a not found: ID does not exist" containerID="9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.926820 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a"} err="failed to get container status \"9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a\": rpc error: code = NotFound desc = could not find container \"9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a\": container with ID starting with 9b4864daf415ad58238396da881fe86492e4b255670c10adfef2650e6127612a not found: ID does not exist" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.926842 4770 scope.go:117] "RemoveContainer" containerID="38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd" Oct 04 05:40:37 crc kubenswrapper[4770]: E1004 05:40:37.927469 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd\": container with ID starting with 38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd not found: ID does not exist" containerID="38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.927525 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd"} err="failed to get container status \"38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd\": rpc error: code = NotFound desc = could not find container \"38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd\": container with ID starting with 38d8c872e29317bd87f0156e1344db30177eb1a4120f5eb51394535bbd417dfd not found: ID does not exist" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.927562 4770 scope.go:117] "RemoveContainer" containerID="3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e" Oct 04 05:40:37 crc kubenswrapper[4770]: E1004 05:40:37.927888 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e\": container with ID starting with 3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e not found: ID does not exist" containerID="3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e" Oct 04 05:40:37 crc kubenswrapper[4770]: I1004 05:40:37.927925 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e"} err="failed to get container status \"3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e\": rpc error: code = NotFound desc = could not find container \"3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e\": container with ID starting with 3ee9365ab13da521f70e77ed9d759560982334f61909b47e47f2174ce3d6687e not found: ID does not exist" Oct 04 05:40:39 crc kubenswrapper[4770]: I1004 05:40:39.693691 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8631a507-43e1-4923-a057-54f930b3fdab" path="/var/lib/kubelet/pods/8631a507-43e1-4923-a057-54f930b3fdab/volumes" Oct 04 05:41:01 crc kubenswrapper[4770]: I1004 05:41:01.796377 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:41:01 crc kubenswrapper[4770]: I1004 05:41:01.797213 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:41:01 crc kubenswrapper[4770]: I1004 05:41:01.797307 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:41:01 crc kubenswrapper[4770]: I1004 05:41:01.798678 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ecefd076d2acf3bcf8b76829f054fa851e70adb6ef01cc6778eaf1cc5cb4ce6"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:41:01 crc kubenswrapper[4770]: I1004 05:41:01.798816 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://3ecefd076d2acf3bcf8b76829f054fa851e70adb6ef01cc6778eaf1cc5cb4ce6" gracePeriod=600 Oct 04 05:41:02 crc kubenswrapper[4770]: I1004 05:41:02.106414 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="3ecefd076d2acf3bcf8b76829f054fa851e70adb6ef01cc6778eaf1cc5cb4ce6" exitCode=0 Oct 04 05:41:02 crc kubenswrapper[4770]: I1004 05:41:02.106917 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"3ecefd076d2acf3bcf8b76829f054fa851e70adb6ef01cc6778eaf1cc5cb4ce6"} Oct 04 05:41:02 crc kubenswrapper[4770]: I1004 05:41:02.107033 4770 scope.go:117] "RemoveContainer" containerID="9450a49a04ed2128c88b261d0123181589ff269eac040ca5bed8da14ef9dc24a" Oct 04 05:41:03 crc kubenswrapper[4770]: I1004 05:41:03.124100 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd"} Oct 04 05:42:44 crc kubenswrapper[4770]: I1004 05:42:44.167060 4770 generic.go:334] "Generic (PLEG): container finished" podID="6707d041-c3e2-4445-b381-ce39c1e572a8" containerID="b38f32eb7dcbd1d9c3e733da028380e28a5891763e81fc9aee49e3f7ce31f59e" exitCode=0 Oct 04 05:42:44 crc kubenswrapper[4770]: I1004 05:42:44.167172 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" event={"ID":"6707d041-c3e2-4445-b381-ce39c1e572a8","Type":"ContainerDied","Data":"b38f32eb7dcbd1d9c3e733da028380e28a5891763e81fc9aee49e3f7ce31f59e"} Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.699796 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.724115 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-inventory\") pod \"6707d041-c3e2-4445-b381-ce39c1e572a8\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.725465 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ssh-key\") pod \"6707d041-c3e2-4445-b381-ce39c1e572a8\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.760297 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6707d041-c3e2-4445-b381-ce39c1e572a8" (UID: "6707d041-c3e2-4445-b381-ce39c1e572a8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.767700 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-inventory" (OuterVolumeSpecName: "inventory") pod "6707d041-c3e2-4445-b381-ce39c1e572a8" (UID: "6707d041-c3e2-4445-b381-ce39c1e572a8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.826941 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbzp2\" (UniqueName: \"kubernetes.io/projected/6707d041-c3e2-4445-b381-ce39c1e572a8-kube-api-access-mbzp2\") pod \"6707d041-c3e2-4445-b381-ce39c1e572a8\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.827031 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-agent-neutron-config-0\") pod \"6707d041-c3e2-4445-b381-ce39c1e572a8\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.827202 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-combined-ca-bundle\") pod \"6707d041-c3e2-4445-b381-ce39c1e572a8\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.827282 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ceph\") pod \"6707d041-c3e2-4445-b381-ce39c1e572a8\" (UID: \"6707d041-c3e2-4445-b381-ce39c1e572a8\") " Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.827641 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.827656 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.831185 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ceph" (OuterVolumeSpecName: "ceph") pod "6707d041-c3e2-4445-b381-ce39c1e572a8" (UID: "6707d041-c3e2-4445-b381-ce39c1e572a8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.831611 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "6707d041-c3e2-4445-b381-ce39c1e572a8" (UID: "6707d041-c3e2-4445-b381-ce39c1e572a8"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.831634 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6707d041-c3e2-4445-b381-ce39c1e572a8-kube-api-access-mbzp2" (OuterVolumeSpecName: "kube-api-access-mbzp2") pod "6707d041-c3e2-4445-b381-ce39c1e572a8" (UID: "6707d041-c3e2-4445-b381-ce39c1e572a8"). InnerVolumeSpecName "kube-api-access-mbzp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.860769 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "6707d041-c3e2-4445-b381-ce39c1e572a8" (UID: "6707d041-c3e2-4445-b381-ce39c1e572a8"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.929855 4770 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.929912 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.929928 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbzp2\" (UniqueName: \"kubernetes.io/projected/6707d041-c3e2-4445-b381-ce39c1e572a8-kube-api-access-mbzp2\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:45 crc kubenswrapper[4770]: I1004 05:42:45.929943 4770 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6707d041-c3e2-4445-b381-ce39c1e572a8-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:46 crc kubenswrapper[4770]: I1004 05:42:46.194120 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" event={"ID":"6707d041-c3e2-4445-b381-ce39c1e572a8","Type":"ContainerDied","Data":"85adb7f2d11a33f98675bd926a8841b5d5750aa04cfe9ccb32eeefa9406aff1d"} Oct 04 05:42:46 crc kubenswrapper[4770]: I1004 05:42:46.194195 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85adb7f2d11a33f98675bd926a8841b5d5750aa04cfe9ccb32eeefa9406aff1d" Oct 04 05:42:46 crc kubenswrapper[4770]: I1004 05:42:46.194634 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-68ssf" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.117375 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.118198 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="6eceb699-4221-46eb-9b62-8ae2cf60cd5b" containerName="nova-cell0-conductor-conductor" containerID="cri-o://bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2" gracePeriod=30 Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.177021 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.177235 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="6db3e1d8-a142-4b4e-b08b-51dc5a2973b0" containerName="nova-cell1-conductor-conductor" containerID="cri-o://608b9cfe609d2917d7ab9daf7942e98dba655ca0c09309610a6944f7b5d12fb7" gracePeriod=30 Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.820760 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw"] Oct 04 05:42:50 crc kubenswrapper[4770]: E1004 05:42:50.821267 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8631a507-43e1-4923-a057-54f930b3fdab" containerName="extract-utilities" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.821285 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8631a507-43e1-4923-a057-54f930b3fdab" containerName="extract-utilities" Oct 04 05:42:50 crc kubenswrapper[4770]: E1004 05:42:50.821303 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8631a507-43e1-4923-a057-54f930b3fdab" containerName="extract-content" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.821309 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8631a507-43e1-4923-a057-54f930b3fdab" containerName="extract-content" Oct 04 05:42:50 crc kubenswrapper[4770]: E1004 05:42:50.821342 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8631a507-43e1-4923-a057-54f930b3fdab" containerName="registry-server" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.821348 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8631a507-43e1-4923-a057-54f930b3fdab" containerName="registry-server" Oct 04 05:42:50 crc kubenswrapper[4770]: E1004 05:42:50.821358 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6707d041-c3e2-4445-b381-ce39c1e572a8" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.821365 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6707d041-c3e2-4445-b381-ce39c1e572a8" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.821605 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8631a507-43e1-4923-a057-54f930b3fdab" containerName="registry-server" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.821620 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6707d041-c3e2-4445-b381-ce39c1e572a8" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.822515 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.825213 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.825477 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.826389 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.826391 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.826539 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.826604 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.826692 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-jxnz5" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.833360 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw"] Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.843893 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.843944 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.844089 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.844176 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws8f6\" (UniqueName: \"kubernetes.io/projected/9549539a-734a-482a-a374-fbd7be4dc5cc-kube-api-access-ws8f6\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.844358 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.844391 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.844426 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.844478 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.844776 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.844823 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.844866 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946106 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946186 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946212 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946247 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946268 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws8f6\" (UniqueName: \"kubernetes.io/projected/9549539a-734a-482a-a374-fbd7be4dc5cc-kube-api-access-ws8f6\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946313 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946335 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946354 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946377 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946449 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.946470 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.947553 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.947766 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.953345 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.953813 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.953921 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.964530 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.965134 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.965235 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.965769 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.969892 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws8f6\" (UniqueName: \"kubernetes.io/projected/9549539a-734a-482a-a374-fbd7be4dc5cc-kube-api-access-ws8f6\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:50 crc kubenswrapper[4770]: I1004 05:42:50.978330 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:51 crc kubenswrapper[4770]: I1004 05:42:51.157166 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:42:51 crc kubenswrapper[4770]: I1004 05:42:51.908669 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw"] Oct 04 05:42:51 crc kubenswrapper[4770]: I1004 05:42:51.941231 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.101122 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.101631 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="46e7a275-1a83-4ee4-bbee-819afd9e1df6" containerName="nova-scheduler-scheduler" containerID="cri-o://a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63" gracePeriod=30 Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.144269 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.144536 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-log" containerID="cri-o://aac27019eb2828d08ce64585131c37f09c142e7a091370cca0513e33c3524167" gracePeriod=30 Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.144990 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-api" containerID="cri-o://67ea243dcc9ab412f13757ffb58194953795759d6551db8bd466f9ad8a7d83ca" gracePeriod=30 Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.170299 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.171067 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-metadata" containerID="cri-o://f60d1eeb54bc17f640f15c972d2bb137d7e0a8e712981795e91597a9ddc441b8" gracePeriod=30 Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.171204 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-log" containerID="cri-o://099d516172bf35064ad73e6e674f0abde980b6c7dba602194a147839a424eddd" gracePeriod=30 Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.281692 4770 generic.go:334] "Generic (PLEG): container finished" podID="6db3e1d8-a142-4b4e-b08b-51dc5a2973b0" containerID="608b9cfe609d2917d7ab9daf7942e98dba655ca0c09309610a6944f7b5d12fb7" exitCode=0 Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.281795 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0","Type":"ContainerDied","Data":"608b9cfe609d2917d7ab9daf7942e98dba655ca0c09309610a6944f7b5d12fb7"} Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.281827 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0","Type":"ContainerDied","Data":"47645a3176d23e317c2a5febfcd448a03051e6e523c6d3ef8e31cef9045111f8"} Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.281842 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47645a3176d23e317c2a5febfcd448a03051e6e523c6d3ef8e31cef9045111f8" Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.285154 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" event={"ID":"9549539a-734a-482a-a374-fbd7be4dc5cc","Type":"ContainerStarted","Data":"42058f536cd0a622d61e30bb028b7b64cacffc512932ce56b144cb09e612c156"} Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.375171 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.487460 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rx49\" (UniqueName: \"kubernetes.io/projected/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-kube-api-access-5rx49\") pod \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.487922 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-combined-ca-bundle\") pod \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.488149 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-config-data\") pod \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\" (UID: \"6db3e1d8-a142-4b4e-b08b-51dc5a2973b0\") " Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.493964 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-kube-api-access-5rx49" (OuterVolumeSpecName: "kube-api-access-5rx49") pod "6db3e1d8-a142-4b4e-b08b-51dc5a2973b0" (UID: "6db3e1d8-a142-4b4e-b08b-51dc5a2973b0"). InnerVolumeSpecName "kube-api-access-5rx49". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.531538 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6db3e1d8-a142-4b4e-b08b-51dc5a2973b0" (UID: "6db3e1d8-a142-4b4e-b08b-51dc5a2973b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.533928 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-config-data" (OuterVolumeSpecName: "config-data") pod "6db3e1d8-a142-4b4e-b08b-51dc5a2973b0" (UID: "6db3e1d8-a142-4b4e-b08b-51dc5a2973b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.592058 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.592091 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rx49\" (UniqueName: \"kubernetes.io/projected/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-kube-api-access-5rx49\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:52 crc kubenswrapper[4770]: I1004 05:42:52.592105 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.344767 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" event={"ID":"9549539a-734a-482a-a374-fbd7be4dc5cc","Type":"ContainerStarted","Data":"e9f3277d28549e2d9783a2fcd1c937d05376654920eed0307d091f77bdd94b86"} Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.349768 4770 generic.go:334] "Generic (PLEG): container finished" podID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerID="099d516172bf35064ad73e6e674f0abde980b6c7dba602194a147839a424eddd" exitCode=143 Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.349845 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b59068af-fca9-4d62-ba15-7f7306a32ae0","Type":"ContainerDied","Data":"099d516172bf35064ad73e6e674f0abde980b6c7dba602194a147839a424eddd"} Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.353133 4770 generic.go:334] "Generic (PLEG): container finished" podID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerID="aac27019eb2828d08ce64585131c37f09c142e7a091370cca0513e33c3524167" exitCode=143 Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.353229 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.354132 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78da1490-9b4b-442a-bb3f-0ad05e1dbf04","Type":"ContainerDied","Data":"aac27019eb2828d08ce64585131c37f09c142e7a091370cca0513e33c3524167"} Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.383356 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" podStartSLOduration=2.857696129 podStartE2EDuration="3.383338694s" podCreationTimestamp="2025-10-04 05:42:50 +0000 UTC" firstStartedPulling="2025-10-04 05:42:51.941022332 +0000 UTC m=+9583.233032044" lastFinishedPulling="2025-10-04 05:42:52.466664897 +0000 UTC m=+9583.758674609" observedRunningTime="2025-10-04 05:42:53.375334968 +0000 UTC m=+9584.667344690" watchObservedRunningTime="2025-10-04 05:42:53.383338694 +0000 UTC m=+9584.675348406" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.488746 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.518015 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.558273 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 05:42:53 crc kubenswrapper[4770]: E1004 05:42:53.558740 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db3e1d8-a142-4b4e-b08b-51dc5a2973b0" containerName="nova-cell1-conductor-conductor" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.558757 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db3e1d8-a142-4b4e-b08b-51dc5a2973b0" containerName="nova-cell1-conductor-conductor" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.559027 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6db3e1d8-a142-4b4e-b08b-51dc5a2973b0" containerName="nova-cell1-conductor-conductor" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.559786 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.574688 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.588190 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.642442 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq6qj\" (UniqueName: \"kubernetes.io/projected/6e87d157-36eb-4197-a066-cefba0bddb76-kube-api-access-mq6qj\") pod \"nova-cell1-conductor-0\" (UID: \"6e87d157-36eb-4197-a066-cefba0bddb76\") " pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.642830 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e87d157-36eb-4197-a066-cefba0bddb76-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6e87d157-36eb-4197-a066-cefba0bddb76\") " pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.642952 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e87d157-36eb-4197-a066-cefba0bddb76-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6e87d157-36eb-4197-a066-cefba0bddb76\") " pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: E1004 05:42:53.666348 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2 is running failed: container process not found" containerID="bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 05:42:53 crc kubenswrapper[4770]: E1004 05:42:53.666775 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2 is running failed: container process not found" containerID="bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 05:42:53 crc kubenswrapper[4770]: E1004 05:42:53.670145 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2 is running failed: container process not found" containerID="bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 04 05:42:53 crc kubenswrapper[4770]: E1004 05:42:53.670355 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="6eceb699-4221-46eb-9b62-8ae2cf60cd5b" containerName="nova-cell0-conductor-conductor" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.697219 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6db3e1d8-a142-4b4e-b08b-51dc5a2973b0" path="/var/lib/kubelet/pods/6db3e1d8-a142-4b4e-b08b-51dc5a2973b0/volumes" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.745278 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e87d157-36eb-4197-a066-cefba0bddb76-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6e87d157-36eb-4197-a066-cefba0bddb76\") " pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.745358 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e87d157-36eb-4197-a066-cefba0bddb76-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6e87d157-36eb-4197-a066-cefba0bddb76\") " pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.746540 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq6qj\" (UniqueName: \"kubernetes.io/projected/6e87d157-36eb-4197-a066-cefba0bddb76-kube-api-access-mq6qj\") pod \"nova-cell1-conductor-0\" (UID: \"6e87d157-36eb-4197-a066-cefba0bddb76\") " pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.756704 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e87d157-36eb-4197-a066-cefba0bddb76-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6e87d157-36eb-4197-a066-cefba0bddb76\") " pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.771704 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq6qj\" (UniqueName: \"kubernetes.io/projected/6e87d157-36eb-4197-a066-cefba0bddb76-kube-api-access-mq6qj\") pod \"nova-cell1-conductor-0\" (UID: \"6e87d157-36eb-4197-a066-cefba0bddb76\") " pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.772198 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e87d157-36eb-4197-a066-cefba0bddb76-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6e87d157-36eb-4197-a066-cefba0bddb76\") " pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:53 crc kubenswrapper[4770]: I1004 05:42:53.944243 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:54 crc kubenswrapper[4770]: E1004 05:42:54.110126 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63 is running failed: container process not found" containerID="a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 05:42:54 crc kubenswrapper[4770]: E1004 05:42:54.110565 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63 is running failed: container process not found" containerID="a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 05:42:54 crc kubenswrapper[4770]: E1004 05:42:54.111071 4770 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63 is running failed: container process not found" containerID="a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 05:42:54 crc kubenswrapper[4770]: E1004 05:42:54.111138 4770 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="46e7a275-1a83-4ee4-bbee-819afd9e1df6" containerName="nova-scheduler-scheduler" Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.386772 4770 generic.go:334] "Generic (PLEG): container finished" podID="6eceb699-4221-46eb-9b62-8ae2cf60cd5b" containerID="bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2" exitCode=0 Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.386850 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"6eceb699-4221-46eb-9b62-8ae2cf60cd5b","Type":"ContainerDied","Data":"bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2"} Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.389228 4770 generic.go:334] "Generic (PLEG): container finished" podID="46e7a275-1a83-4ee4-bbee-819afd9e1df6" containerID="a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63" exitCode=0 Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.390107 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"46e7a275-1a83-4ee4-bbee-819afd9e1df6","Type":"ContainerDied","Data":"a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63"} Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.477870 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 05:42:54 crc kubenswrapper[4770]: W1004 05:42:54.486635 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e87d157_36eb_4197_a066_cefba0bddb76.slice/crio-5f32d690a8f01505b1f71c48b2383dbb7ed35acdea0daaee48e31dc59a7f4e6a WatchSource:0}: Error finding container 5f32d690a8f01505b1f71c48b2383dbb7ed35acdea0daaee48e31dc59a7f4e6a: Status 404 returned error can't find the container with id 5f32d690a8f01505b1f71c48b2383dbb7ed35acdea0daaee48e31dc59a7f4e6a Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.679513 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.775319 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-config-data\") pod \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.775456 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-combined-ca-bundle\") pod \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.775556 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzscp\" (UniqueName: \"kubernetes.io/projected/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-kube-api-access-tzscp\") pod \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\" (UID: \"6eceb699-4221-46eb-9b62-8ae2cf60cd5b\") " Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.787609 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-kube-api-access-tzscp" (OuterVolumeSpecName: "kube-api-access-tzscp") pod "6eceb699-4221-46eb-9b62-8ae2cf60cd5b" (UID: "6eceb699-4221-46eb-9b62-8ae2cf60cd5b"). InnerVolumeSpecName "kube-api-access-tzscp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.827082 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-config-data" (OuterVolumeSpecName: "config-data") pod "6eceb699-4221-46eb-9b62-8ae2cf60cd5b" (UID: "6eceb699-4221-46eb-9b62-8ae2cf60cd5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.828033 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6eceb699-4221-46eb-9b62-8ae2cf60cd5b" (UID: "6eceb699-4221-46eb-9b62-8ae2cf60cd5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.878440 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.878479 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.878494 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzscp\" (UniqueName: \"kubernetes.io/projected/6eceb699-4221-46eb-9b62-8ae2cf60cd5b-kube-api-access-tzscp\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.925251 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.980103 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64h79\" (UniqueName: \"kubernetes.io/projected/46e7a275-1a83-4ee4-bbee-819afd9e1df6-kube-api-access-64h79\") pod \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.980190 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-config-data\") pod \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.980380 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-combined-ca-bundle\") pod \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\" (UID: \"46e7a275-1a83-4ee4-bbee-819afd9e1df6\") " Oct 04 05:42:54 crc kubenswrapper[4770]: I1004 05:42:54.983666 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46e7a275-1a83-4ee4-bbee-819afd9e1df6-kube-api-access-64h79" (OuterVolumeSpecName: "kube-api-access-64h79") pod "46e7a275-1a83-4ee4-bbee-819afd9e1df6" (UID: "46e7a275-1a83-4ee4-bbee-819afd9e1df6"). InnerVolumeSpecName "kube-api-access-64h79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.010539 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46e7a275-1a83-4ee4-bbee-819afd9e1df6" (UID: "46e7a275-1a83-4ee4-bbee-819afd9e1df6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.016821 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-config-data" (OuterVolumeSpecName: "config-data") pod "46e7a275-1a83-4ee4-bbee-819afd9e1df6" (UID: "46e7a275-1a83-4ee4-bbee-819afd9e1df6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.083491 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.083528 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64h79\" (UniqueName: \"kubernetes.io/projected/46e7a275-1a83-4ee4-bbee-819afd9e1df6-kube-api-access-64h79\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.083544 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46e7a275-1a83-4ee4-bbee-819afd9e1df6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.399499 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"6eceb699-4221-46eb-9b62-8ae2cf60cd5b","Type":"ContainerDied","Data":"d0c0ad0ee77b6bd2213ee3ef438a3b31557fc30be183f588241e17f90c732e40"} Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.399839 4770 scope.go:117] "RemoveContainer" containerID="bf32323d0d66d292f38916deee1e43f1628597e0f247cca2dc4a99ccb6d945f2" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.399528 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.401085 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6e87d157-36eb-4197-a066-cefba0bddb76","Type":"ContainerStarted","Data":"8ce188db5d164018eb169085e6a99e474fef09dafea259674ae90db12302ee87"} Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.401128 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6e87d157-36eb-4197-a066-cefba0bddb76","Type":"ContainerStarted","Data":"5f32d690a8f01505b1f71c48b2383dbb7ed35acdea0daaee48e31dc59a7f4e6a"} Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.401525 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.402893 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"46e7a275-1a83-4ee4-bbee-819afd9e1df6","Type":"ContainerDied","Data":"55d139908fc565872075917de4d0a292f7ea2528f96553b97e4dba033641718b"} Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.402940 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.435091 4770 scope.go:117] "RemoveContainer" containerID="a633c615801b09315d5ffce04627ef3f569cd11fe7d01710986b82d2f52c6c63" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.437103 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.4370908399999998 podStartE2EDuration="2.43709084s" podCreationTimestamp="2025-10-04 05:42:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:42:55.427341647 +0000 UTC m=+9586.719351359" watchObservedRunningTime="2025-10-04 05:42:55.43709084 +0000 UTC m=+9586.729100552" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.453233 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.468326 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.487592 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.515693 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.524344 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 05:42:55 crc kubenswrapper[4770]: E1004 05:42:55.524800 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46e7a275-1a83-4ee4-bbee-819afd9e1df6" containerName="nova-scheduler-scheduler" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.524822 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="46e7a275-1a83-4ee4-bbee-819afd9e1df6" containerName="nova-scheduler-scheduler" Oct 04 05:42:55 crc kubenswrapper[4770]: E1004 05:42:55.524845 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eceb699-4221-46eb-9b62-8ae2cf60cd5b" containerName="nova-cell0-conductor-conductor" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.524854 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eceb699-4221-46eb-9b62-8ae2cf60cd5b" containerName="nova-cell0-conductor-conductor" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.525146 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eceb699-4221-46eb-9b62-8ae2cf60cd5b" containerName="nova-cell0-conductor-conductor" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.525171 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="46e7a275-1a83-4ee4-bbee-819afd9e1df6" containerName="nova-scheduler-scheduler" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.525927 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.528393 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.534509 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.542501 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.544372 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.546252 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.554079 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.593395 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72r5p\" (UniqueName: \"kubernetes.io/projected/ca0f0db0-e939-4054-9aa4-8d610293082f-kube-api-access-72r5p\") pod \"nova-scheduler-0\" (UID: \"ca0f0db0-e939-4054-9aa4-8d610293082f\") " pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.593446 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca0f0db0-e939-4054-9aa4-8d610293082f-config-data\") pod \"nova-scheduler-0\" (UID: \"ca0f0db0-e939-4054-9aa4-8d610293082f\") " pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.593494 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28662e31-5054-4536-84c7-e22941f7dc19-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"28662e31-5054-4536-84c7-e22941f7dc19\") " pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.593513 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca0f0db0-e939-4054-9aa4-8d610293082f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca0f0db0-e939-4054-9aa4-8d610293082f\") " pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.593544 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28662e31-5054-4536-84c7-e22941f7dc19-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"28662e31-5054-4536-84c7-e22941f7dc19\") " pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.593588 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7g5t\" (UniqueName: \"kubernetes.io/projected/28662e31-5054-4536-84c7-e22941f7dc19-kube-api-access-b7g5t\") pod \"nova-cell0-conductor-0\" (UID: \"28662e31-5054-4536-84c7-e22941f7dc19\") " pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.647105 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": read tcp 10.217.0.2:48518->10.217.1.84:8775: read: connection reset by peer" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.647139 4770 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": read tcp 10.217.0.2:48526->10.217.1.84:8775: read: connection reset by peer" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.693554 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46e7a275-1a83-4ee4-bbee-819afd9e1df6" path="/var/lib/kubelet/pods/46e7a275-1a83-4ee4-bbee-819afd9e1df6/volumes" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.695151 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eceb699-4221-46eb-9b62-8ae2cf60cd5b" path="/var/lib/kubelet/pods/6eceb699-4221-46eb-9b62-8ae2cf60cd5b/volumes" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.698586 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72r5p\" (UniqueName: \"kubernetes.io/projected/ca0f0db0-e939-4054-9aa4-8d610293082f-kube-api-access-72r5p\") pod \"nova-scheduler-0\" (UID: \"ca0f0db0-e939-4054-9aa4-8d610293082f\") " pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.698655 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca0f0db0-e939-4054-9aa4-8d610293082f-config-data\") pod \"nova-scheduler-0\" (UID: \"ca0f0db0-e939-4054-9aa4-8d610293082f\") " pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.698697 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28662e31-5054-4536-84c7-e22941f7dc19-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"28662e31-5054-4536-84c7-e22941f7dc19\") " pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.698727 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca0f0db0-e939-4054-9aa4-8d610293082f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca0f0db0-e939-4054-9aa4-8d610293082f\") " pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.698771 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28662e31-5054-4536-84c7-e22941f7dc19-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"28662e31-5054-4536-84c7-e22941f7dc19\") " pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.698829 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7g5t\" (UniqueName: \"kubernetes.io/projected/28662e31-5054-4536-84c7-e22941f7dc19-kube-api-access-b7g5t\") pod \"nova-cell0-conductor-0\" (UID: \"28662e31-5054-4536-84c7-e22941f7dc19\") " pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.706199 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28662e31-5054-4536-84c7-e22941f7dc19-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"28662e31-5054-4536-84c7-e22941f7dc19\") " pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.710842 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca0f0db0-e939-4054-9aa4-8d610293082f-config-data\") pod \"nova-scheduler-0\" (UID: \"ca0f0db0-e939-4054-9aa4-8d610293082f\") " pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.713746 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28662e31-5054-4536-84c7-e22941f7dc19-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"28662e31-5054-4536-84c7-e22941f7dc19\") " pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.718110 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca0f0db0-e939-4054-9aa4-8d610293082f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca0f0db0-e939-4054-9aa4-8d610293082f\") " pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.718330 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72r5p\" (UniqueName: \"kubernetes.io/projected/ca0f0db0-e939-4054-9aa4-8d610293082f-kube-api-access-72r5p\") pod \"nova-scheduler-0\" (UID: \"ca0f0db0-e939-4054-9aa4-8d610293082f\") " pod="openstack/nova-scheduler-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.730451 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7g5t\" (UniqueName: \"kubernetes.io/projected/28662e31-5054-4536-84c7-e22941f7dc19-kube-api-access-b7g5t\") pod \"nova-cell0-conductor-0\" (UID: \"28662e31-5054-4536-84c7-e22941f7dc19\") " pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.914538 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:55 crc kubenswrapper[4770]: I1004 05:42:55.929708 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.435495 4770 generic.go:334] "Generic (PLEG): container finished" podID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerID="f60d1eeb54bc17f640f15c972d2bb137d7e0a8e712981795e91597a9ddc441b8" exitCode=0 Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.435728 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b59068af-fca9-4d62-ba15-7f7306a32ae0","Type":"ContainerDied","Data":"f60d1eeb54bc17f640f15c972d2bb137d7e0a8e712981795e91597a9ddc441b8"} Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.444525 4770 generic.go:334] "Generic (PLEG): container finished" podID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerID="67ea243dcc9ab412f13757ffb58194953795759d6551db8bd466f9ad8a7d83ca" exitCode=0 Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.444607 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78da1490-9b4b-442a-bb3f-0ad05e1dbf04","Type":"ContainerDied","Data":"67ea243dcc9ab412f13757ffb58194953795759d6551db8bd466f9ad8a7d83ca"} Oct 04 05:42:56 crc kubenswrapper[4770]: W1004 05:42:56.457214 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28662e31_5054_4536_84c7_e22941f7dc19.slice/crio-00564d69d8d9bee1457864b99baa09c4e6b58c053e83cfca8f2f57f785747598 WatchSource:0}: Error finding container 00564d69d8d9bee1457864b99baa09c4e6b58c053e83cfca8f2f57f785747598: Status 404 returned error can't find the container with id 00564d69d8d9bee1457864b99baa09c4e6b58c053e83cfca8f2f57f785747598 Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.463658 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.564484 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.884697 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.951841 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-combined-ca-bundle\") pod \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.952363 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-logs\") pod \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.952407 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9czd\" (UniqueName: \"kubernetes.io/projected/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-kube-api-access-t9czd\") pod \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.952573 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-config-data\") pod \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.952904 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-logs" (OuterVolumeSpecName: "logs") pod "78da1490-9b4b-442a-bb3f-0ad05e1dbf04" (UID: "78da1490-9b4b-442a-bb3f-0ad05e1dbf04"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.953982 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-logs\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:56 crc kubenswrapper[4770]: I1004 05:42:56.958222 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-kube-api-access-t9czd" (OuterVolumeSpecName: "kube-api-access-t9czd") pod "78da1490-9b4b-442a-bb3f-0ad05e1dbf04" (UID: "78da1490-9b4b-442a-bb3f-0ad05e1dbf04"). InnerVolumeSpecName "kube-api-access-t9czd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:42:56 crc kubenswrapper[4770]: E1004 05:42:56.996438 4770 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-config-data podName:78da1490-9b4b-442a-bb3f-0ad05e1dbf04 nodeName:}" failed. No retries permitted until 2025-10-04 05:42:57.496407351 +0000 UTC m=+9588.788417063 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-config-data") pod "78da1490-9b4b-442a-bb3f-0ad05e1dbf04" (UID: "78da1490-9b4b-442a-bb3f-0ad05e1dbf04") : error deleting /var/lib/kubelet/pods/78da1490-9b4b-442a-bb3f-0ad05e1dbf04/volume-subpaths: remove /var/lib/kubelet/pods/78da1490-9b4b-442a-bb3f-0ad05e1dbf04/volume-subpaths: no such file or directory Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.003252 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78da1490-9b4b-442a-bb3f-0ad05e1dbf04" (UID: "78da1490-9b4b-442a-bb3f-0ad05e1dbf04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.015185 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.055345 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxzvl\" (UniqueName: \"kubernetes.io/projected/b59068af-fca9-4d62-ba15-7f7306a32ae0-kube-api-access-fxzvl\") pod \"b59068af-fca9-4d62-ba15-7f7306a32ae0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.055486 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59068af-fca9-4d62-ba15-7f7306a32ae0-logs\") pod \"b59068af-fca9-4d62-ba15-7f7306a32ae0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.055603 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-config-data\") pod \"b59068af-fca9-4d62-ba15-7f7306a32ae0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.055728 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-combined-ca-bundle\") pod \"b59068af-fca9-4d62-ba15-7f7306a32ae0\" (UID: \"b59068af-fca9-4d62-ba15-7f7306a32ae0\") " Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.056575 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9czd\" (UniqueName: \"kubernetes.io/projected/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-kube-api-access-t9czd\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.056593 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.057826 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59068af-fca9-4d62-ba15-7f7306a32ae0-logs" (OuterVolumeSpecName: "logs") pod "b59068af-fca9-4d62-ba15-7f7306a32ae0" (UID: "b59068af-fca9-4d62-ba15-7f7306a32ae0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.064276 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59068af-fca9-4d62-ba15-7f7306a32ae0-kube-api-access-fxzvl" (OuterVolumeSpecName: "kube-api-access-fxzvl") pod "b59068af-fca9-4d62-ba15-7f7306a32ae0" (UID: "b59068af-fca9-4d62-ba15-7f7306a32ae0"). InnerVolumeSpecName "kube-api-access-fxzvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.103694 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-config-data" (OuterVolumeSpecName: "config-data") pod "b59068af-fca9-4d62-ba15-7f7306a32ae0" (UID: "b59068af-fca9-4d62-ba15-7f7306a32ae0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.107922 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b59068af-fca9-4d62-ba15-7f7306a32ae0" (UID: "b59068af-fca9-4d62-ba15-7f7306a32ae0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.161164 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxzvl\" (UniqueName: \"kubernetes.io/projected/b59068af-fca9-4d62-ba15-7f7306a32ae0-kube-api-access-fxzvl\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.161194 4770 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59068af-fca9-4d62-ba15-7f7306a32ae0-logs\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.161205 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.161213 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59068af-fca9-4d62-ba15-7f7306a32ae0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.463163 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca0f0db0-e939-4054-9aa4-8d610293082f","Type":"ContainerStarted","Data":"b771542bffee315f2dd4e5c84219a180a8f36335f68ddb2e89f407558f243070"} Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.464201 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca0f0db0-e939-4054-9aa4-8d610293082f","Type":"ContainerStarted","Data":"2f067e00bed923f070cdd1ba3b79088f71d0ce1ba6b744f4057bc3ba57767158"} Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.468326 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"28662e31-5054-4536-84c7-e22941f7dc19","Type":"ContainerStarted","Data":"88a02b4b81c5b4037a663da8de4f2239eede8babdd7b97484bc7c3887459f9fd"} Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.468367 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"28662e31-5054-4536-84c7-e22941f7dc19","Type":"ContainerStarted","Data":"00564d69d8d9bee1457864b99baa09c4e6b58c053e83cfca8f2f57f785747598"} Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.471367 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b59068af-fca9-4d62-ba15-7f7306a32ae0","Type":"ContainerDied","Data":"7b695814a33d8874da6ca761c52f4b75ac4e102287f87d7f15e3455a90af64eb"} Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.471649 4770 scope.go:117] "RemoveContainer" containerID="f60d1eeb54bc17f640f15c972d2bb137d7e0a8e712981795e91597a9ddc441b8" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.471612 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.474506 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78da1490-9b4b-442a-bb3f-0ad05e1dbf04","Type":"ContainerDied","Data":"1367a090f87a30ea7101871c1d7d58d4a3ca79f66dbf43f1908fae718d3bae28"} Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.474614 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.499608 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.499591081 podStartE2EDuration="2.499591081s" podCreationTimestamp="2025-10-04 05:42:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:42:57.488301336 +0000 UTC m=+9588.780311058" watchObservedRunningTime="2025-10-04 05:42:57.499591081 +0000 UTC m=+9588.791600793" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.522250 4770 scope.go:117] "RemoveContainer" containerID="099d516172bf35064ad73e6e674f0abde980b6c7dba602194a147839a424eddd" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.559348 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.559330813 podStartE2EDuration="2.559330813s" podCreationTimestamp="2025-10-04 05:42:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:42:57.542311774 +0000 UTC m=+9588.834321496" watchObservedRunningTime="2025-10-04 05:42:57.559330813 +0000 UTC m=+9588.851340525" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.569722 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-config-data\") pod \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\" (UID: \"78da1490-9b4b-442a-bb3f-0ad05e1dbf04\") " Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.571841 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.582380 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.582455 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-config-data" (OuterVolumeSpecName: "config-data") pod "78da1490-9b4b-442a-bb3f-0ad05e1dbf04" (UID: "78da1490-9b4b-442a-bb3f-0ad05e1dbf04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.591813 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:42:57 crc kubenswrapper[4770]: E1004 05:42:57.592709 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-api" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.592732 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-api" Oct 04 05:42:57 crc kubenswrapper[4770]: E1004 05:42:57.592767 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-log" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.592777 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-log" Oct 04 05:42:57 crc kubenswrapper[4770]: E1004 05:42:57.592815 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-log" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.592824 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-log" Oct 04 05:42:57 crc kubenswrapper[4770]: E1004 05:42:57.592840 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-metadata" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.592848 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-metadata" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.593156 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-metadata" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.593178 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-api" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.593214 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" containerName="nova-api-log" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.593243 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" containerName="nova-metadata-log" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.594860 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.597590 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.601874 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.673220 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-config-data\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.673294 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-logs\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.673328 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnflh\" (UniqueName: \"kubernetes.io/projected/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-kube-api-access-tnflh\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.674039 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.674243 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78da1490-9b4b-442a-bb3f-0ad05e1dbf04-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.692460 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b59068af-fca9-4d62-ba15-7f7306a32ae0" path="/var/lib/kubelet/pods/b59068af-fca9-4d62-ba15-7f7306a32ae0/volumes" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.711124 4770 scope.go:117] "RemoveContainer" containerID="67ea243dcc9ab412f13757ffb58194953795759d6551db8bd466f9ad8a7d83ca" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.754601 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pqlb7"] Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.762711 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.768270 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pqlb7"] Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.784678 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-config-data\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.784739 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-logs\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.784768 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnflh\" (UniqueName: \"kubernetes.io/projected/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-kube-api-access-tnflh\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.784825 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.786427 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-logs\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.790753 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.806814 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-config-data\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.829486 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnflh\" (UniqueName: \"kubernetes.io/projected/c3ce320f-ae54-4e20-915d-c1c1a9a1c887-kube-api-access-tnflh\") pod \"nova-metadata-0\" (UID: \"c3ce320f-ae54-4e20-915d-c1c1a9a1c887\") " pod="openstack/nova-metadata-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.835559 4770 scope.go:117] "RemoveContainer" containerID="aac27019eb2828d08ce64585131c37f09c142e7a091370cca0513e33c3524167" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.887574 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlfhw\" (UniqueName: \"kubernetes.io/projected/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-kube-api-access-vlfhw\") pod \"certified-operators-pqlb7\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.887902 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-catalog-content\") pod \"certified-operators-pqlb7\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.888248 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-utilities\") pod \"certified-operators-pqlb7\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.925562 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.941481 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.965390 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.967697 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.975889 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.979633 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.990096 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlfhw\" (UniqueName: \"kubernetes.io/projected/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-kube-api-access-vlfhw\") pod \"certified-operators-pqlb7\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.990167 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-catalog-content\") pod \"certified-operators-pqlb7\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.990344 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-utilities\") pod \"certified-operators-pqlb7\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.990741 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-utilities\") pod \"certified-operators-pqlb7\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:57 crc kubenswrapper[4770]: I1004 05:42:57.990800 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-catalog-content\") pod \"certified-operators-pqlb7\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.035825 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlfhw\" (UniqueName: \"kubernetes.io/projected/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-kube-api-access-vlfhw\") pod \"certified-operators-pqlb7\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.048232 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.098807 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbff40a9-a564-4416-8e59-3cc50cef4b90-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.099064 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hl9s\" (UniqueName: \"kubernetes.io/projected/fbff40a9-a564-4416-8e59-3cc50cef4b90-kube-api-access-6hl9s\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.099216 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbff40a9-a564-4416-8e59-3cc50cef4b90-config-data\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.099257 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbff40a9-a564-4416-8e59-3cc50cef4b90-logs\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.200989 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.249283 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hl9s\" (UniqueName: \"kubernetes.io/projected/fbff40a9-a564-4416-8e59-3cc50cef4b90-kube-api-access-6hl9s\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.249696 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbff40a9-a564-4416-8e59-3cc50cef4b90-config-data\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.249724 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbff40a9-a564-4416-8e59-3cc50cef4b90-logs\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.249799 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbff40a9-a564-4416-8e59-3cc50cef4b90-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.251159 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbff40a9-a564-4416-8e59-3cc50cef4b90-logs\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.267913 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbff40a9-a564-4416-8e59-3cc50cef4b90-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.268606 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbff40a9-a564-4416-8e59-3cc50cef4b90-config-data\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.282674 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hl9s\" (UniqueName: \"kubernetes.io/projected/fbff40a9-a564-4416-8e59-3cc50cef4b90-kube-api-access-6hl9s\") pod \"nova-api-0\" (UID: \"fbff40a9-a564-4416-8e59-3cc50cef4b90\") " pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.297534 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.489985 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 05:42:58 crc kubenswrapper[4770]: I1004 05:42:58.866325 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 05:42:58 crc kubenswrapper[4770]: W1004 05:42:58.869916 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3ce320f_ae54_4e20_915d_c1c1a9a1c887.slice/crio-9e22dcc2271a35c6f81960517027d1960c4c8176d7d3faddbb519af0322a5326 WatchSource:0}: Error finding container 9e22dcc2271a35c6f81960517027d1960c4c8176d7d3faddbb519af0322a5326: Status 404 returned error can't find the container with id 9e22dcc2271a35c6f81960517027d1960c4c8176d7d3faddbb519af0322a5326 Oct 04 05:42:59 crc kubenswrapper[4770]: I1004 05:42:59.371557 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 05:42:59 crc kubenswrapper[4770]: I1004 05:42:59.501427 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbff40a9-a564-4416-8e59-3cc50cef4b90","Type":"ContainerStarted","Data":"f1912b82fc30eea8001406eaaa13521a782be31a3f96cb82baa349bfc30e5f91"} Oct 04 05:42:59 crc kubenswrapper[4770]: I1004 05:42:59.506440 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3ce320f-ae54-4e20-915d-c1c1a9a1c887","Type":"ContainerStarted","Data":"7106ef3e6c49bf016d883db45067a15e5647a96600cd11bbb2da1e79c72b73be"} Oct 04 05:42:59 crc kubenswrapper[4770]: I1004 05:42:59.506483 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3ce320f-ae54-4e20-915d-c1c1a9a1c887","Type":"ContainerStarted","Data":"9e22dcc2271a35c6f81960517027d1960c4c8176d7d3faddbb519af0322a5326"} Oct 04 05:42:59 crc kubenswrapper[4770]: I1004 05:42:59.533086 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pqlb7"] Oct 04 05:42:59 crc kubenswrapper[4770]: I1004 05:42:59.701065 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78da1490-9b4b-442a-bb3f-0ad05e1dbf04" path="/var/lib/kubelet/pods/78da1490-9b4b-442a-bb3f-0ad05e1dbf04/volumes" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.515966 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p5zqq"] Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.520117 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.520623 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3ce320f-ae54-4e20-915d-c1c1a9a1c887","Type":"ContainerStarted","Data":"d290b1cfaf090f6d9a6c1572a3cede4f98b92a28e5ec4e6b1758987ec3ab4bc0"} Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.526501 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbff40a9-a564-4416-8e59-3cc50cef4b90","Type":"ContainerStarted","Data":"3f8cb63728eef68a94a0ab7c93223ae9dc23c8235a7a3fc53f0c91da1a34266d"} Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.526714 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbff40a9-a564-4416-8e59-3cc50cef4b90","Type":"ContainerStarted","Data":"d03363015fda88ced3bf5e6fdeb0e722106feb98800403377d7a43a58b7342b8"} Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.529939 4770 generic.go:334] "Generic (PLEG): container finished" podID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerID="5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83" exitCode=0 Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.529992 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqlb7" event={"ID":"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85","Type":"ContainerDied","Data":"5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83"} Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.530036 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqlb7" event={"ID":"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85","Type":"ContainerStarted","Data":"810af8691d6a378e9ac41bbac1b25e61c938132400a16d5af246975c61f60a34"} Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.530407 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5zqq"] Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.577029 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.576989861 podStartE2EDuration="3.576989861s" podCreationTimestamp="2025-10-04 05:42:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:43:00.538997905 +0000 UTC m=+9591.831007627" watchObservedRunningTime="2025-10-04 05:43:00.576989861 +0000 UTC m=+9591.868999583" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.609867 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-catalog-content\") pod \"community-operators-p5zqq\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.610078 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-utilities\") pod \"community-operators-p5zqq\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.610150 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ptz6\" (UniqueName: \"kubernetes.io/projected/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-kube-api-access-2ptz6\") pod \"community-operators-p5zqq\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.618738 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.618715257 podStartE2EDuration="3.618715257s" podCreationTimestamp="2025-10-04 05:42:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:43:00.590575117 +0000 UTC m=+9591.882584839" watchObservedRunningTime="2025-10-04 05:43:00.618715257 +0000 UTC m=+9591.910724969" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.711766 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-catalog-content\") pod \"community-operators-p5zqq\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.711918 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-utilities\") pod \"community-operators-p5zqq\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.711970 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ptz6\" (UniqueName: \"kubernetes.io/projected/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-kube-api-access-2ptz6\") pod \"community-operators-p5zqq\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.712495 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-utilities\") pod \"community-operators-p5zqq\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.712643 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-catalog-content\") pod \"community-operators-p5zqq\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.930538 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 05:43:00 crc kubenswrapper[4770]: I1004 05:43:00.944298 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ptz6\" (UniqueName: \"kubernetes.io/projected/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-kube-api-access-2ptz6\") pod \"community-operators-p5zqq\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:01 crc kubenswrapper[4770]: I1004 05:43:01.148449 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:01 crc kubenswrapper[4770]: I1004 05:43:01.669745 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5zqq"] Oct 04 05:43:02 crc kubenswrapper[4770]: I1004 05:43:02.553419 4770 generic.go:334] "Generic (PLEG): container finished" podID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerID="db8fab8576993c1f59e3d24d04eb1f06b439f299c56b83448f7283ac19157d65" exitCode=0 Oct 04 05:43:02 crc kubenswrapper[4770]: I1004 05:43:02.553507 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5zqq" event={"ID":"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4","Type":"ContainerDied","Data":"db8fab8576993c1f59e3d24d04eb1f06b439f299c56b83448f7283ac19157d65"} Oct 04 05:43:02 crc kubenswrapper[4770]: I1004 05:43:02.554209 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5zqq" event={"ID":"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4","Type":"ContainerStarted","Data":"fe0a661edda9b175eaafe0f0df48adc7fc4c1dde4c87f76f600c1b9864812cd6"} Oct 04 05:43:03 crc kubenswrapper[4770]: I1004 05:43:03.049765 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 05:43:03 crc kubenswrapper[4770]: I1004 05:43:03.050170 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 05:43:03 crc kubenswrapper[4770]: I1004 05:43:03.566859 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqlb7" event={"ID":"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85","Type":"ContainerStarted","Data":"572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1"} Oct 04 05:43:03 crc kubenswrapper[4770]: I1004 05:43:03.976217 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 05:43:05 crc kubenswrapper[4770]: I1004 05:43:05.605139 4770 generic.go:334] "Generic (PLEG): container finished" podID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerID="11a162c1fc82f04532bdda68d3e3ed134becb6a766d841d2dce66c7fdc22df60" exitCode=0 Oct 04 05:43:05 crc kubenswrapper[4770]: I1004 05:43:05.605797 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5zqq" event={"ID":"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4","Type":"ContainerDied","Data":"11a162c1fc82f04532bdda68d3e3ed134becb6a766d841d2dce66c7fdc22df60"} Oct 04 05:43:05 crc kubenswrapper[4770]: I1004 05:43:05.930999 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 05:43:05 crc kubenswrapper[4770]: I1004 05:43:05.951235 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 05:43:05 crc kubenswrapper[4770]: I1004 05:43:05.971909 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 05:43:06 crc kubenswrapper[4770]: I1004 05:43:06.620587 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5zqq" event={"ID":"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4","Type":"ContainerStarted","Data":"4f6406cf9140d7fd58a5ddcfdfd971a4137a6b5b995a569232f4288c82854c40"} Oct 04 05:43:06 crc kubenswrapper[4770]: I1004 05:43:06.623367 4770 generic.go:334] "Generic (PLEG): container finished" podID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerID="572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1" exitCode=0 Oct 04 05:43:06 crc kubenswrapper[4770]: I1004 05:43:06.624582 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqlb7" event={"ID":"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85","Type":"ContainerDied","Data":"572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1"} Oct 04 05:43:06 crc kubenswrapper[4770]: I1004 05:43:06.669413 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 05:43:07 crc kubenswrapper[4770]: I1004 05:43:07.639483 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqlb7" event={"ID":"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85","Type":"ContainerStarted","Data":"8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2"} Oct 04 05:43:07 crc kubenswrapper[4770]: I1004 05:43:07.677929 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p5zqq" podStartSLOduration=4.076341447 podStartE2EDuration="7.677910192s" podCreationTimestamp="2025-10-04 05:43:00 +0000 UTC" firstStartedPulling="2025-10-04 05:43:02.555988717 +0000 UTC m=+9593.847998429" lastFinishedPulling="2025-10-04 05:43:06.157557462 +0000 UTC m=+9597.449567174" observedRunningTime="2025-10-04 05:43:07.664458899 +0000 UTC m=+9598.956468621" watchObservedRunningTime="2025-10-04 05:43:07.677910192 +0000 UTC m=+9598.969919894" Oct 04 05:43:07 crc kubenswrapper[4770]: I1004 05:43:07.704702 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pqlb7" podStartSLOduration=4.056440479 podStartE2EDuration="10.704679515s" podCreationTimestamp="2025-10-04 05:42:57 +0000 UTC" firstStartedPulling="2025-10-04 05:43:00.533836105 +0000 UTC m=+9591.825845817" lastFinishedPulling="2025-10-04 05:43:07.182075141 +0000 UTC m=+9598.474084853" observedRunningTime="2025-10-04 05:43:07.695641711 +0000 UTC m=+9598.987651423" watchObservedRunningTime="2025-10-04 05:43:07.704679515 +0000 UTC m=+9598.996689227" Oct 04 05:43:08 crc kubenswrapper[4770]: I1004 05:43:08.049921 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 05:43:08 crc kubenswrapper[4770]: I1004 05:43:08.050315 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 05:43:08 crc kubenswrapper[4770]: I1004 05:43:08.211180 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:43:08 crc kubenswrapper[4770]: I1004 05:43:08.211230 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:43:08 crc kubenswrapper[4770]: I1004 05:43:08.298976 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 05:43:08 crc kubenswrapper[4770]: I1004 05:43:08.299038 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 05:43:09 crc kubenswrapper[4770]: I1004 05:43:09.090389 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3ce320f-ae54-4e20-915d-c1c1a9a1c887" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.196:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 05:43:09 crc kubenswrapper[4770]: I1004 05:43:09.131396 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3ce320f-ae54-4e20-915d-c1c1a9a1c887" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.196:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 05:43:09 crc kubenswrapper[4770]: I1004 05:43:09.269588 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-pqlb7" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerName="registry-server" probeResult="failure" output=< Oct 04 05:43:09 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:43:09 crc kubenswrapper[4770]: > Oct 04 05:43:09 crc kubenswrapper[4770]: I1004 05:43:09.381251 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fbff40a9-a564-4416-8e59-3cc50cef4b90" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 05:43:09 crc kubenswrapper[4770]: I1004 05:43:09.381607 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fbff40a9-a564-4416-8e59-3cc50cef4b90" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 05:43:11 crc kubenswrapper[4770]: I1004 05:43:11.149158 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:11 crc kubenswrapper[4770]: I1004 05:43:11.149674 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:11 crc kubenswrapper[4770]: I1004 05:43:11.200201 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:11 crc kubenswrapper[4770]: I1004 05:43:11.745383 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:12 crc kubenswrapper[4770]: I1004 05:43:12.309311 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5zqq"] Oct 04 05:43:13 crc kubenswrapper[4770]: I1004 05:43:13.706781 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p5zqq" podUID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerName="registry-server" containerID="cri-o://4f6406cf9140d7fd58a5ddcfdfd971a4137a6b5b995a569232f4288c82854c40" gracePeriod=2 Oct 04 05:43:14 crc kubenswrapper[4770]: I1004 05:43:14.723414 4770 generic.go:334] "Generic (PLEG): container finished" podID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerID="4f6406cf9140d7fd58a5ddcfdfd971a4137a6b5b995a569232f4288c82854c40" exitCode=0 Oct 04 05:43:14 crc kubenswrapper[4770]: I1004 05:43:14.723717 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5zqq" event={"ID":"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4","Type":"ContainerDied","Data":"4f6406cf9140d7fd58a5ddcfdfd971a4137a6b5b995a569232f4288c82854c40"} Oct 04 05:43:14 crc kubenswrapper[4770]: I1004 05:43:14.990279 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.064734 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ptz6\" (UniqueName: \"kubernetes.io/projected/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-kube-api-access-2ptz6\") pod \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.065023 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-catalog-content\") pod \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.065090 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-utilities\") pod \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\" (UID: \"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4\") " Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.066292 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-utilities" (OuterVolumeSpecName: "utilities") pod "f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" (UID: "f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.070114 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-kube-api-access-2ptz6" (OuterVolumeSpecName: "kube-api-access-2ptz6") pod "f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" (UID: "f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4"). InnerVolumeSpecName "kube-api-access-2ptz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.121095 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" (UID: "f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.168071 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.168117 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.168131 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ptz6\" (UniqueName: \"kubernetes.io/projected/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4-kube-api-access-2ptz6\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.739424 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5zqq" event={"ID":"f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4","Type":"ContainerDied","Data":"fe0a661edda9b175eaafe0f0df48adc7fc4c1dde4c87f76f600c1b9864812cd6"} Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.739488 4770 scope.go:117] "RemoveContainer" containerID="4f6406cf9140d7fd58a5ddcfdfd971a4137a6b5b995a569232f4288c82854c40" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.739495 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5zqq" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.771057 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5zqq"] Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.774685 4770 scope.go:117] "RemoveContainer" containerID="11a162c1fc82f04532bdda68d3e3ed134becb6a766d841d2dce66c7fdc22df60" Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.784957 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p5zqq"] Oct 04 05:43:15 crc kubenswrapper[4770]: I1004 05:43:15.805744 4770 scope.go:117] "RemoveContainer" containerID="db8fab8576993c1f59e3d24d04eb1f06b439f299c56b83448f7283ac19157d65" Oct 04 05:43:17 crc kubenswrapper[4770]: I1004 05:43:17.687604 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" path="/var/lib/kubelet/pods/f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4/volumes" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.051494 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.054195 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.055872 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.257303 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.302755 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.303129 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.303429 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.308703 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.312616 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.705374 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pqlb7"] Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.768679 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.771848 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 05:43:18 crc kubenswrapper[4770]: I1004 05:43:18.772390 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 05:43:19 crc kubenswrapper[4770]: I1004 05:43:19.777736 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pqlb7" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerName="registry-server" containerID="cri-o://8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2" gracePeriod=2 Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.388705 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.505132 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-catalog-content\") pod \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.505241 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlfhw\" (UniqueName: \"kubernetes.io/projected/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-kube-api-access-vlfhw\") pod \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.505369 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-utilities\") pod \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\" (UID: \"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85\") " Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.506553 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-utilities" (OuterVolumeSpecName: "utilities") pod "3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" (UID: "3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.517895 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-kube-api-access-vlfhw" (OuterVolumeSpecName: "kube-api-access-vlfhw") pod "3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" (UID: "3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85"). InnerVolumeSpecName "kube-api-access-vlfhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.553213 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" (UID: "3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.608323 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.608363 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlfhw\" (UniqueName: \"kubernetes.io/projected/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-kube-api-access-vlfhw\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.608373 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.791710 4770 generic.go:334] "Generic (PLEG): container finished" podID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerID="8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2" exitCode=0 Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.791791 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqlb7" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.791790 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqlb7" event={"ID":"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85","Type":"ContainerDied","Data":"8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2"} Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.791869 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqlb7" event={"ID":"3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85","Type":"ContainerDied","Data":"810af8691d6a378e9ac41bbac1b25e61c938132400a16d5af246975c61f60a34"} Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.791929 4770 scope.go:117] "RemoveContainer" containerID="8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.825378 4770 scope.go:117] "RemoveContainer" containerID="572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.829190 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pqlb7"] Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.840056 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pqlb7"] Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.866850 4770 scope.go:117] "RemoveContainer" containerID="5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.905545 4770 scope.go:117] "RemoveContainer" containerID="8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2" Oct 04 05:43:20 crc kubenswrapper[4770]: E1004 05:43:20.906169 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2\": container with ID starting with 8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2 not found: ID does not exist" containerID="8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.906203 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2"} err="failed to get container status \"8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2\": rpc error: code = NotFound desc = could not find container \"8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2\": container with ID starting with 8b78a8244810800fd6790bc1d22bbbd9001b069f4f26280466d3ec674fbed4f2 not found: ID does not exist" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.906240 4770 scope.go:117] "RemoveContainer" containerID="572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1" Oct 04 05:43:20 crc kubenswrapper[4770]: E1004 05:43:20.906730 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1\": container with ID starting with 572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1 not found: ID does not exist" containerID="572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.906777 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1"} err="failed to get container status \"572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1\": rpc error: code = NotFound desc = could not find container \"572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1\": container with ID starting with 572985e977a4e1795a84b71f7412a92ad8975e85d70a08919dbb97b634f15bb1 not found: ID does not exist" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.906806 4770 scope.go:117] "RemoveContainer" containerID="5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83" Oct 04 05:43:20 crc kubenswrapper[4770]: E1004 05:43:20.907157 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83\": container with ID starting with 5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83 not found: ID does not exist" containerID="5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83" Oct 04 05:43:20 crc kubenswrapper[4770]: I1004 05:43:20.907261 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83"} err="failed to get container status \"5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83\": rpc error: code = NotFound desc = could not find container \"5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83\": container with ID starting with 5b15c029970bc259003a86aab908f259762c6b75464ab3ccf700ebf2d7c52f83 not found: ID does not exist" Oct 04 05:43:21 crc kubenswrapper[4770]: I1004 05:43:21.686363 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" path="/var/lib/kubelet/pods/3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85/volumes" Oct 04 05:43:31 crc kubenswrapper[4770]: I1004 05:43:31.796048 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:43:31 crc kubenswrapper[4770]: I1004 05:43:31.798723 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:43:32 crc kubenswrapper[4770]: I1004 05:43:32.499967 4770 scope.go:117] "RemoveContainer" containerID="608b9cfe609d2917d7ab9daf7942e98dba655ca0c09309610a6944f7b5d12fb7" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.690672 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4mdgm"] Oct 04 05:43:37 crc kubenswrapper[4770]: E1004 05:43:37.691757 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerName="extract-content" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.691773 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerName="extract-content" Oct 04 05:43:37 crc kubenswrapper[4770]: E1004 05:43:37.691794 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerName="registry-server" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.691803 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerName="registry-server" Oct 04 05:43:37 crc kubenswrapper[4770]: E1004 05:43:37.691819 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerName="extract-utilities" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.691825 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerName="extract-utilities" Oct 04 05:43:37 crc kubenswrapper[4770]: E1004 05:43:37.691833 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerName="extract-utilities" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.691839 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerName="extract-utilities" Oct 04 05:43:37 crc kubenswrapper[4770]: E1004 05:43:37.691849 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerName="registry-server" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.691854 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerName="registry-server" Oct 04 05:43:37 crc kubenswrapper[4770]: E1004 05:43:37.691873 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerName="extract-content" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.691879 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerName="extract-content" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.692113 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b177fdc-18b1-4b2e-99ba-0d4ec6bdde85" containerName="registry-server" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.692131 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7486ca3-cff2-4e9d-9054-ddb88c7f7cd4" containerName="registry-server" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.694220 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.697340 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4mdgm"] Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.827761 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-catalog-content\") pod \"redhat-operators-4mdgm\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.830828 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-utilities\") pod \"redhat-operators-4mdgm\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.830967 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzksz\" (UniqueName: \"kubernetes.io/projected/61399c39-3a0f-4498-aa02-2b0cd2652a9c-kube-api-access-jzksz\") pod \"redhat-operators-4mdgm\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.932374 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-utilities\") pod \"redhat-operators-4mdgm\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.932550 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzksz\" (UniqueName: \"kubernetes.io/projected/61399c39-3a0f-4498-aa02-2b0cd2652a9c-kube-api-access-jzksz\") pod \"redhat-operators-4mdgm\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.932597 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-catalog-content\") pod \"redhat-operators-4mdgm\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.946489 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-utilities\") pod \"redhat-operators-4mdgm\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.947595 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-catalog-content\") pod \"redhat-operators-4mdgm\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:37 crc kubenswrapper[4770]: I1004 05:43:37.957122 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzksz\" (UniqueName: \"kubernetes.io/projected/61399c39-3a0f-4498-aa02-2b0cd2652a9c-kube-api-access-jzksz\") pod \"redhat-operators-4mdgm\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:38 crc kubenswrapper[4770]: I1004 05:43:38.026359 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:43:38 crc kubenswrapper[4770]: I1004 05:43:38.629868 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4mdgm"] Oct 04 05:43:38 crc kubenswrapper[4770]: I1004 05:43:38.993721 4770 generic.go:334] "Generic (PLEG): container finished" podID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerID="0414ffabb15ffbe11cd2a6f0b4885924c1991bbc32945c994c7f506b31d7d1c0" exitCode=0 Oct 04 05:43:38 crc kubenswrapper[4770]: I1004 05:43:38.993771 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mdgm" event={"ID":"61399c39-3a0f-4498-aa02-2b0cd2652a9c","Type":"ContainerDied","Data":"0414ffabb15ffbe11cd2a6f0b4885924c1991bbc32945c994c7f506b31d7d1c0"} Oct 04 05:43:38 crc kubenswrapper[4770]: I1004 05:43:38.993799 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mdgm" event={"ID":"61399c39-3a0f-4498-aa02-2b0cd2652a9c","Type":"ContainerStarted","Data":"06a731953e1bfd05360d0e164c9a1e96c5d9a29da1058d19079d9801af79ab10"} Oct 04 05:43:41 crc kubenswrapper[4770]: I1004 05:43:41.014059 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mdgm" event={"ID":"61399c39-3a0f-4498-aa02-2b0cd2652a9c","Type":"ContainerStarted","Data":"a7a42d25077c829b991e4bdab07459f8493d1a9ac3d48bfd042c54ea368b1a45"} Oct 04 05:43:58 crc kubenswrapper[4770]: I1004 05:43:58.224255 4770 generic.go:334] "Generic (PLEG): container finished" podID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerID="a7a42d25077c829b991e4bdab07459f8493d1a9ac3d48bfd042c54ea368b1a45" exitCode=0 Oct 04 05:43:58 crc kubenswrapper[4770]: I1004 05:43:58.224318 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mdgm" event={"ID":"61399c39-3a0f-4498-aa02-2b0cd2652a9c","Type":"ContainerDied","Data":"a7a42d25077c829b991e4bdab07459f8493d1a9ac3d48bfd042c54ea368b1a45"} Oct 04 05:44:00 crc kubenswrapper[4770]: I1004 05:44:00.247855 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mdgm" event={"ID":"61399c39-3a0f-4498-aa02-2b0cd2652a9c","Type":"ContainerStarted","Data":"c194d40d9dd97ddba989820a5fa47cc14ea8a42c8f88864d96779111a5888a27"} Oct 04 05:44:00 crc kubenswrapper[4770]: I1004 05:44:00.276690 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4mdgm" podStartSLOduration=2.663827694 podStartE2EDuration="23.276663128s" podCreationTimestamp="2025-10-04 05:43:37 +0000 UTC" firstStartedPulling="2025-10-04 05:43:38.997296064 +0000 UTC m=+9630.289305776" lastFinishedPulling="2025-10-04 05:43:59.610131498 +0000 UTC m=+9650.902141210" observedRunningTime="2025-10-04 05:44:00.27154164 +0000 UTC m=+9651.563551352" watchObservedRunningTime="2025-10-04 05:44:00.276663128 +0000 UTC m=+9651.568672860" Oct 04 05:44:01 crc kubenswrapper[4770]: I1004 05:44:01.796104 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:44:01 crc kubenswrapper[4770]: I1004 05:44:01.796491 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:44:08 crc kubenswrapper[4770]: I1004 05:44:08.027446 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:44:08 crc kubenswrapper[4770]: I1004 05:44:08.028133 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:44:08 crc kubenswrapper[4770]: I1004 05:44:08.077111 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:44:08 crc kubenswrapper[4770]: I1004 05:44:08.991220 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:44:09 crc kubenswrapper[4770]: I1004 05:44:09.054220 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4mdgm"] Oct 04 05:44:10 crc kubenswrapper[4770]: I1004 05:44:10.345153 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4mdgm" podUID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerName="registry-server" containerID="cri-o://c194d40d9dd97ddba989820a5fa47cc14ea8a42c8f88864d96779111a5888a27" gracePeriod=2 Oct 04 05:44:11 crc kubenswrapper[4770]: I1004 05:44:11.363352 4770 generic.go:334] "Generic (PLEG): container finished" podID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerID="c194d40d9dd97ddba989820a5fa47cc14ea8a42c8f88864d96779111a5888a27" exitCode=0 Oct 04 05:44:11 crc kubenswrapper[4770]: I1004 05:44:11.363714 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mdgm" event={"ID":"61399c39-3a0f-4498-aa02-2b0cd2652a9c","Type":"ContainerDied","Data":"c194d40d9dd97ddba989820a5fa47cc14ea8a42c8f88864d96779111a5888a27"} Oct 04 05:44:11 crc kubenswrapper[4770]: I1004 05:44:11.893819 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:44:11 crc kubenswrapper[4770]: I1004 05:44:11.993479 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzksz\" (UniqueName: \"kubernetes.io/projected/61399c39-3a0f-4498-aa02-2b0cd2652a9c-kube-api-access-jzksz\") pod \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " Oct 04 05:44:11 crc kubenswrapper[4770]: I1004 05:44:11.993620 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-catalog-content\") pod \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " Oct 04 05:44:11 crc kubenswrapper[4770]: I1004 05:44:11.993686 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-utilities\") pod \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\" (UID: \"61399c39-3a0f-4498-aa02-2b0cd2652a9c\") " Oct 04 05:44:11 crc kubenswrapper[4770]: I1004 05:44:11.994521 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-utilities" (OuterVolumeSpecName: "utilities") pod "61399c39-3a0f-4498-aa02-2b0cd2652a9c" (UID: "61399c39-3a0f-4498-aa02-2b0cd2652a9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.036330 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61399c39-3a0f-4498-aa02-2b0cd2652a9c-kube-api-access-jzksz" (OuterVolumeSpecName: "kube-api-access-jzksz") pod "61399c39-3a0f-4498-aa02-2b0cd2652a9c" (UID: "61399c39-3a0f-4498-aa02-2b0cd2652a9c"). InnerVolumeSpecName "kube-api-access-jzksz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.081644 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61399c39-3a0f-4498-aa02-2b0cd2652a9c" (UID: "61399c39-3a0f-4498-aa02-2b0cd2652a9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.096218 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzksz\" (UniqueName: \"kubernetes.io/projected/61399c39-3a0f-4498-aa02-2b0cd2652a9c-kube-api-access-jzksz\") on node \"crc\" DevicePath \"\"" Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.096257 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.096272 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61399c39-3a0f-4498-aa02-2b0cd2652a9c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.376567 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4mdgm" event={"ID":"61399c39-3a0f-4498-aa02-2b0cd2652a9c","Type":"ContainerDied","Data":"06a731953e1bfd05360d0e164c9a1e96c5d9a29da1058d19079d9801af79ab10"} Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.376657 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4mdgm" Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.376941 4770 scope.go:117] "RemoveContainer" containerID="c194d40d9dd97ddba989820a5fa47cc14ea8a42c8f88864d96779111a5888a27" Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.420609 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4mdgm"] Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.421358 4770 scope.go:117] "RemoveContainer" containerID="a7a42d25077c829b991e4bdab07459f8493d1a9ac3d48bfd042c54ea368b1a45" Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.433740 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4mdgm"] Oct 04 05:44:12 crc kubenswrapper[4770]: I1004 05:44:12.456129 4770 scope.go:117] "RemoveContainer" containerID="0414ffabb15ffbe11cd2a6f0b4885924c1991bbc32945c994c7f506b31d7d1c0" Oct 04 05:44:13 crc kubenswrapper[4770]: I1004 05:44:13.688257 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" path="/var/lib/kubelet/pods/61399c39-3a0f-4498-aa02-2b0cd2652a9c/volumes" Oct 04 05:44:31 crc kubenswrapper[4770]: I1004 05:44:31.795640 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:44:31 crc kubenswrapper[4770]: I1004 05:44:31.796251 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:44:31 crc kubenswrapper[4770]: I1004 05:44:31.796306 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:44:31 crc kubenswrapper[4770]: I1004 05:44:31.797324 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:44:31 crc kubenswrapper[4770]: I1004 05:44:31.797392 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" gracePeriod=600 Oct 04 05:44:32 crc kubenswrapper[4770]: I1004 05:44:32.608923 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" exitCode=0 Oct 04 05:44:32 crc kubenswrapper[4770]: I1004 05:44:32.609639 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd"} Oct 04 05:44:32 crc kubenswrapper[4770]: I1004 05:44:32.609741 4770 scope.go:117] "RemoveContainer" containerID="3ecefd076d2acf3bcf8b76829f054fa851e70adb6ef01cc6778eaf1cc5cb4ce6" Oct 04 05:44:32 crc kubenswrapper[4770]: E1004 05:44:32.866532 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:44:33 crc kubenswrapper[4770]: I1004 05:44:33.625643 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:44:33 crc kubenswrapper[4770]: E1004 05:44:33.626337 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:44:45 crc kubenswrapper[4770]: I1004 05:44:45.674559 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:44:45 crc kubenswrapper[4770]: E1004 05:44:45.675465 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:44:59 crc kubenswrapper[4770]: I1004 05:44:59.685112 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:44:59 crc kubenswrapper[4770]: E1004 05:44:59.686029 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.168816 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5"] Oct 04 05:45:00 crc kubenswrapper[4770]: E1004 05:45:00.169348 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerName="extract-utilities" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.169371 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerName="extract-utilities" Oct 04 05:45:00 crc kubenswrapper[4770]: E1004 05:45:00.169394 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerName="extract-content" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.169403 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerName="extract-content" Oct 04 05:45:00 crc kubenswrapper[4770]: E1004 05:45:00.169419 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.169426 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.169694 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="61399c39-3a0f-4498-aa02-2b0cd2652a9c" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.170516 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.173241 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.173259 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.183079 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5"] Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.308322 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8591141-2b4d-4f20-9981-6dcb39ae3698-secret-volume\") pod \"collect-profiles-29325945-n5sk5\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.308897 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dm9w\" (UniqueName: \"kubernetes.io/projected/c8591141-2b4d-4f20-9981-6dcb39ae3698-kube-api-access-7dm9w\") pod \"collect-profiles-29325945-n5sk5\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.309100 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8591141-2b4d-4f20-9981-6dcb39ae3698-config-volume\") pod \"collect-profiles-29325945-n5sk5\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.411387 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8591141-2b4d-4f20-9981-6dcb39ae3698-config-volume\") pod \"collect-profiles-29325945-n5sk5\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.411522 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8591141-2b4d-4f20-9981-6dcb39ae3698-secret-volume\") pod \"collect-profiles-29325945-n5sk5\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.411593 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dm9w\" (UniqueName: \"kubernetes.io/projected/c8591141-2b4d-4f20-9981-6dcb39ae3698-kube-api-access-7dm9w\") pod \"collect-profiles-29325945-n5sk5\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.412846 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8591141-2b4d-4f20-9981-6dcb39ae3698-config-volume\") pod \"collect-profiles-29325945-n5sk5\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.418308 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8591141-2b4d-4f20-9981-6dcb39ae3698-secret-volume\") pod \"collect-profiles-29325945-n5sk5\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.430575 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dm9w\" (UniqueName: \"kubernetes.io/projected/c8591141-2b4d-4f20-9981-6dcb39ae3698-kube-api-access-7dm9w\") pod \"collect-profiles-29325945-n5sk5\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.501044 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:00 crc kubenswrapper[4770]: I1004 05:45:00.984718 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5"] Oct 04 05:45:01 crc kubenswrapper[4770]: I1004 05:45:01.971382 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" event={"ID":"c8591141-2b4d-4f20-9981-6dcb39ae3698","Type":"ContainerStarted","Data":"722d7084bf535235ebb6cb0b1b540aefcf301e1b2f760aff16b44beabe1ac4cd"} Oct 04 05:45:01 crc kubenswrapper[4770]: I1004 05:45:01.971882 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" event={"ID":"c8591141-2b4d-4f20-9981-6dcb39ae3698","Type":"ContainerStarted","Data":"df6220536de20c6f4084c30fb9a98953f02d29ac4606312246a828630e6852a7"} Oct 04 05:45:01 crc kubenswrapper[4770]: I1004 05:45:01.996290 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" podStartSLOduration=1.996272488 podStartE2EDuration="1.996272488s" podCreationTimestamp="2025-10-04 05:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:45:01.991683874 +0000 UTC m=+9713.283693596" watchObservedRunningTime="2025-10-04 05:45:01.996272488 +0000 UTC m=+9713.288282210" Oct 04 05:45:02 crc kubenswrapper[4770]: I1004 05:45:02.982418 4770 generic.go:334] "Generic (PLEG): container finished" podID="c8591141-2b4d-4f20-9981-6dcb39ae3698" containerID="722d7084bf535235ebb6cb0b1b540aefcf301e1b2f760aff16b44beabe1ac4cd" exitCode=0 Oct 04 05:45:02 crc kubenswrapper[4770]: I1004 05:45:02.982481 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" event={"ID":"c8591141-2b4d-4f20-9981-6dcb39ae3698","Type":"ContainerDied","Data":"722d7084bf535235ebb6cb0b1b540aefcf301e1b2f760aff16b44beabe1ac4cd"} Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.577360 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.620126 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8591141-2b4d-4f20-9981-6dcb39ae3698-secret-volume\") pod \"c8591141-2b4d-4f20-9981-6dcb39ae3698\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.620340 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dm9w\" (UniqueName: \"kubernetes.io/projected/c8591141-2b4d-4f20-9981-6dcb39ae3698-kube-api-access-7dm9w\") pod \"c8591141-2b4d-4f20-9981-6dcb39ae3698\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.620369 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8591141-2b4d-4f20-9981-6dcb39ae3698-config-volume\") pod \"c8591141-2b4d-4f20-9981-6dcb39ae3698\" (UID: \"c8591141-2b4d-4f20-9981-6dcb39ae3698\") " Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.621266 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8591141-2b4d-4f20-9981-6dcb39ae3698-config-volume" (OuterVolumeSpecName: "config-volume") pod "c8591141-2b4d-4f20-9981-6dcb39ae3698" (UID: "c8591141-2b4d-4f20-9981-6dcb39ae3698"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.628660 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8591141-2b4d-4f20-9981-6dcb39ae3698-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c8591141-2b4d-4f20-9981-6dcb39ae3698" (UID: "c8591141-2b4d-4f20-9981-6dcb39ae3698"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.628899 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8591141-2b4d-4f20-9981-6dcb39ae3698-kube-api-access-7dm9w" (OuterVolumeSpecName: "kube-api-access-7dm9w") pod "c8591141-2b4d-4f20-9981-6dcb39ae3698" (UID: "c8591141-2b4d-4f20-9981-6dcb39ae3698"). InnerVolumeSpecName "kube-api-access-7dm9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.724542 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dm9w\" (UniqueName: \"kubernetes.io/projected/c8591141-2b4d-4f20-9981-6dcb39ae3698-kube-api-access-7dm9w\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.724872 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8591141-2b4d-4f20-9981-6dcb39ae3698-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:04 crc kubenswrapper[4770]: I1004 05:45:04.724963 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8591141-2b4d-4f20-9981-6dcb39ae3698-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:05 crc kubenswrapper[4770]: I1004 05:45:05.004261 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" event={"ID":"c8591141-2b4d-4f20-9981-6dcb39ae3698","Type":"ContainerDied","Data":"df6220536de20c6f4084c30fb9a98953f02d29ac4606312246a828630e6852a7"} Oct 04 05:45:05 crc kubenswrapper[4770]: I1004 05:45:05.004738 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df6220536de20c6f4084c30fb9a98953f02d29ac4606312246a828630e6852a7" Oct 04 05:45:05 crc kubenswrapper[4770]: I1004 05:45:05.004308 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-n5sk5" Oct 04 05:45:05 crc kubenswrapper[4770]: I1004 05:45:05.056421 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs"] Oct 04 05:45:05 crc kubenswrapper[4770]: I1004 05:45:05.065041 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zjdfs"] Oct 04 05:45:05 crc kubenswrapper[4770]: I1004 05:45:05.690672 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2274a58e-0eb7-459a-8cf9-568c0f6ca8b5" path="/var/lib/kubelet/pods/2274a58e-0eb7-459a-8cf9-568c0f6ca8b5/volumes" Oct 04 05:45:14 crc kubenswrapper[4770]: I1004 05:45:14.673974 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:45:14 crc kubenswrapper[4770]: E1004 05:45:14.674773 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:45:27 crc kubenswrapper[4770]: I1004 05:45:27.673993 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:45:27 crc kubenswrapper[4770]: E1004 05:45:27.675898 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:45:32 crc kubenswrapper[4770]: I1004 05:45:32.728636 4770 scope.go:117] "RemoveContainer" containerID="0b4cb397514a2569fc50577730957c6ded11b44120b1cea50fae539c8b096c0a" Oct 04 05:45:42 crc kubenswrapper[4770]: I1004 05:45:42.677913 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:45:42 crc kubenswrapper[4770]: E1004 05:45:42.678740 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:45:55 crc kubenswrapper[4770]: I1004 05:45:55.674085 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:45:55 crc kubenswrapper[4770]: E1004 05:45:55.675011 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:46:06 crc kubenswrapper[4770]: I1004 05:46:06.674412 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:46:06 crc kubenswrapper[4770]: E1004 05:46:06.675483 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:46:18 crc kubenswrapper[4770]: I1004 05:46:18.673446 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:46:18 crc kubenswrapper[4770]: E1004 05:46:18.674254 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:46:31 crc kubenswrapper[4770]: I1004 05:46:31.673955 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:46:31 crc kubenswrapper[4770]: E1004 05:46:31.674828 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:46:46 crc kubenswrapper[4770]: I1004 05:46:46.674342 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:46:46 crc kubenswrapper[4770]: E1004 05:46:46.675971 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:46:58 crc kubenswrapper[4770]: I1004 05:46:58.674578 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:46:58 crc kubenswrapper[4770]: E1004 05:46:58.675441 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:47:10 crc kubenswrapper[4770]: I1004 05:47:10.673801 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:47:10 crc kubenswrapper[4770]: E1004 05:47:10.674671 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:47:23 crc kubenswrapper[4770]: I1004 05:47:23.673807 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:47:23 crc kubenswrapper[4770]: E1004 05:47:23.675436 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:47:37 crc kubenswrapper[4770]: I1004 05:47:37.673343 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:47:37 crc kubenswrapper[4770]: E1004 05:47:37.675111 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:47:49 crc kubenswrapper[4770]: I1004 05:47:49.688275 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:47:49 crc kubenswrapper[4770]: E1004 05:47:49.689250 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:48:03 crc kubenswrapper[4770]: I1004 05:48:03.674709 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:48:03 crc kubenswrapper[4770]: E1004 05:48:03.675523 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:48:14 crc kubenswrapper[4770]: I1004 05:48:14.673792 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:48:14 crc kubenswrapper[4770]: E1004 05:48:14.674876 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:48:27 crc kubenswrapper[4770]: I1004 05:48:27.674159 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:48:27 crc kubenswrapper[4770]: E1004 05:48:27.675033 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:48:39 crc kubenswrapper[4770]: I1004 05:48:39.681322 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:48:39 crc kubenswrapper[4770]: E1004 05:48:39.682349 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:48:51 crc kubenswrapper[4770]: I1004 05:48:51.675351 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:48:51 crc kubenswrapper[4770]: E1004 05:48:51.676146 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:49:02 crc kubenswrapper[4770]: I1004 05:49:02.674058 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:49:02 crc kubenswrapper[4770]: E1004 05:49:02.674993 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:49:17 crc kubenswrapper[4770]: I1004 05:49:17.674246 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:49:17 crc kubenswrapper[4770]: E1004 05:49:17.675152 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:49:32 crc kubenswrapper[4770]: I1004 05:49:32.673551 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:49:33 crc kubenswrapper[4770]: I1004 05:49:33.707808 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"ef7d94b51b87f586c30b02e12461a333be72c1e193d7af0e3b2c10208c30e361"} Oct 04 05:51:43 crc kubenswrapper[4770]: I1004 05:51:43.983988 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wndvh"] Oct 04 05:51:43 crc kubenswrapper[4770]: E1004 05:51:43.985097 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8591141-2b4d-4f20-9981-6dcb39ae3698" containerName="collect-profiles" Oct 04 05:51:43 crc kubenswrapper[4770]: I1004 05:51:43.985113 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8591141-2b4d-4f20-9981-6dcb39ae3698" containerName="collect-profiles" Oct 04 05:51:43 crc kubenswrapper[4770]: I1004 05:51:43.985366 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8591141-2b4d-4f20-9981-6dcb39ae3698" containerName="collect-profiles" Oct 04 05:51:43 crc kubenswrapper[4770]: I1004 05:51:43.989792 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:43 crc kubenswrapper[4770]: I1004 05:51:43.999290 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndvh"] Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.086602 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgx4g\" (UniqueName: \"kubernetes.io/projected/999af072-bf92-44b5-9a4e-4199579ab678-kube-api-access-dgx4g\") pod \"redhat-marketplace-wndvh\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.086659 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-utilities\") pod \"redhat-marketplace-wndvh\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.086700 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-catalog-content\") pod \"redhat-marketplace-wndvh\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.189397 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgx4g\" (UniqueName: \"kubernetes.io/projected/999af072-bf92-44b5-9a4e-4199579ab678-kube-api-access-dgx4g\") pod \"redhat-marketplace-wndvh\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.189464 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-utilities\") pod \"redhat-marketplace-wndvh\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.189508 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-catalog-content\") pod \"redhat-marketplace-wndvh\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.190521 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-catalog-content\") pod \"redhat-marketplace-wndvh\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.190609 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-utilities\") pod \"redhat-marketplace-wndvh\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.214789 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgx4g\" (UniqueName: \"kubernetes.io/projected/999af072-bf92-44b5-9a4e-4199579ab678-kube-api-access-dgx4g\") pod \"redhat-marketplace-wndvh\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.323070 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:44 crc kubenswrapper[4770]: I1004 05:51:44.825525 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndvh"] Oct 04 05:51:45 crc kubenswrapper[4770]: I1004 05:51:45.057378 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndvh" event={"ID":"999af072-bf92-44b5-9a4e-4199579ab678","Type":"ContainerStarted","Data":"8e195288c11f66957b7c0658794bedd5ac6cf18ed6e46cb51c8222b9589b63db"} Oct 04 05:51:46 crc kubenswrapper[4770]: I1004 05:51:46.075268 4770 generic.go:334] "Generic (PLEG): container finished" podID="999af072-bf92-44b5-9a4e-4199579ab678" containerID="b3bfbd846852f5483dfcf2ece69d4dba8d9538659a69a0bf93fbbc1a1703e9c4" exitCode=0 Oct 04 05:51:46 crc kubenswrapper[4770]: I1004 05:51:46.075323 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndvh" event={"ID":"999af072-bf92-44b5-9a4e-4199579ab678","Type":"ContainerDied","Data":"b3bfbd846852f5483dfcf2ece69d4dba8d9538659a69a0bf93fbbc1a1703e9c4"} Oct 04 05:51:46 crc kubenswrapper[4770]: I1004 05:51:46.077778 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:51:47 crc kubenswrapper[4770]: I1004 05:51:47.111810 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndvh" event={"ID":"999af072-bf92-44b5-9a4e-4199579ab678","Type":"ContainerStarted","Data":"0d959a0a32daa533b83736a33dded49df466d181851d1d8fd754fbdd428a0d40"} Oct 04 05:51:48 crc kubenswrapper[4770]: I1004 05:51:48.133397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndvh" event={"ID":"999af072-bf92-44b5-9a4e-4199579ab678","Type":"ContainerDied","Data":"0d959a0a32daa533b83736a33dded49df466d181851d1d8fd754fbdd428a0d40"} Oct 04 05:51:48 crc kubenswrapper[4770]: I1004 05:51:48.133259 4770 generic.go:334] "Generic (PLEG): container finished" podID="999af072-bf92-44b5-9a4e-4199579ab678" containerID="0d959a0a32daa533b83736a33dded49df466d181851d1d8fd754fbdd428a0d40" exitCode=0 Oct 04 05:51:49 crc kubenswrapper[4770]: I1004 05:51:49.149647 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndvh" event={"ID":"999af072-bf92-44b5-9a4e-4199579ab678","Type":"ContainerStarted","Data":"2811eac606421840ccedb2add041de9aef4b138602bde1356bf433e5d33122b2"} Oct 04 05:51:49 crc kubenswrapper[4770]: I1004 05:51:49.173922 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wndvh" podStartSLOduration=3.702579577 podStartE2EDuration="6.17390234s" podCreationTimestamp="2025-10-04 05:51:43 +0000 UTC" firstStartedPulling="2025-10-04 05:51:46.077526377 +0000 UTC m=+10117.369536089" lastFinishedPulling="2025-10-04 05:51:48.54884913 +0000 UTC m=+10119.840858852" observedRunningTime="2025-10-04 05:51:49.171252939 +0000 UTC m=+10120.463262651" watchObservedRunningTime="2025-10-04 05:51:49.17390234 +0000 UTC m=+10120.465912042" Oct 04 05:51:54 crc kubenswrapper[4770]: I1004 05:51:54.323529 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:54 crc kubenswrapper[4770]: I1004 05:51:54.324308 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:54 crc kubenswrapper[4770]: I1004 05:51:54.384991 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:55 crc kubenswrapper[4770]: I1004 05:51:55.260798 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:55 crc kubenswrapper[4770]: I1004 05:51:55.315612 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndvh"] Oct 04 05:51:57 crc kubenswrapper[4770]: I1004 05:51:57.233295 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wndvh" podUID="999af072-bf92-44b5-9a4e-4199579ab678" containerName="registry-server" containerID="cri-o://2811eac606421840ccedb2add041de9aef4b138602bde1356bf433e5d33122b2" gracePeriod=2 Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.244927 4770 generic.go:334] "Generic (PLEG): container finished" podID="999af072-bf92-44b5-9a4e-4199579ab678" containerID="2811eac606421840ccedb2add041de9aef4b138602bde1356bf433e5d33122b2" exitCode=0 Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.245233 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndvh" event={"ID":"999af072-bf92-44b5-9a4e-4199579ab678","Type":"ContainerDied","Data":"2811eac606421840ccedb2add041de9aef4b138602bde1356bf433e5d33122b2"} Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.398748 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.591305 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgx4g\" (UniqueName: \"kubernetes.io/projected/999af072-bf92-44b5-9a4e-4199579ab678-kube-api-access-dgx4g\") pod \"999af072-bf92-44b5-9a4e-4199579ab678\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.591693 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-catalog-content\") pod \"999af072-bf92-44b5-9a4e-4199579ab678\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.595923 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-utilities\") pod \"999af072-bf92-44b5-9a4e-4199579ab678\" (UID: \"999af072-bf92-44b5-9a4e-4199579ab678\") " Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.596945 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-utilities" (OuterVolumeSpecName: "utilities") pod "999af072-bf92-44b5-9a4e-4199579ab678" (UID: "999af072-bf92-44b5-9a4e-4199579ab678"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.598882 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/999af072-bf92-44b5-9a4e-4199579ab678-kube-api-access-dgx4g" (OuterVolumeSpecName: "kube-api-access-dgx4g") pod "999af072-bf92-44b5-9a4e-4199579ab678" (UID: "999af072-bf92-44b5-9a4e-4199579ab678"). InnerVolumeSpecName "kube-api-access-dgx4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.604568 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "999af072-bf92-44b5-9a4e-4199579ab678" (UID: "999af072-bf92-44b5-9a4e-4199579ab678"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.698326 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.698375 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgx4g\" (UniqueName: \"kubernetes.io/projected/999af072-bf92-44b5-9a4e-4199579ab678-kube-api-access-dgx4g\") on node \"crc\" DevicePath \"\"" Oct 04 05:51:58 crc kubenswrapper[4770]: I1004 05:51:58.698393 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/999af072-bf92-44b5-9a4e-4199579ab678-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:51:59 crc kubenswrapper[4770]: I1004 05:51:59.257789 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndvh" event={"ID":"999af072-bf92-44b5-9a4e-4199579ab678","Type":"ContainerDied","Data":"8e195288c11f66957b7c0658794bedd5ac6cf18ed6e46cb51c8222b9589b63db"} Oct 04 05:51:59 crc kubenswrapper[4770]: I1004 05:51:59.257852 4770 scope.go:117] "RemoveContainer" containerID="2811eac606421840ccedb2add041de9aef4b138602bde1356bf433e5d33122b2" Oct 04 05:51:59 crc kubenswrapper[4770]: I1004 05:51:59.257854 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wndvh" Oct 04 05:51:59 crc kubenswrapper[4770]: I1004 05:51:59.285349 4770 scope.go:117] "RemoveContainer" containerID="0d959a0a32daa533b83736a33dded49df466d181851d1d8fd754fbdd428a0d40" Oct 04 05:51:59 crc kubenswrapper[4770]: I1004 05:51:59.294190 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndvh"] Oct 04 05:51:59 crc kubenswrapper[4770]: I1004 05:51:59.309137 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndvh"] Oct 04 05:51:59 crc kubenswrapper[4770]: I1004 05:51:59.326446 4770 scope.go:117] "RemoveContainer" containerID="b3bfbd846852f5483dfcf2ece69d4dba8d9538659a69a0bf93fbbc1a1703e9c4" Oct 04 05:51:59 crc kubenswrapper[4770]: I1004 05:51:59.686615 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="999af072-bf92-44b5-9a4e-4199579ab678" path="/var/lib/kubelet/pods/999af072-bf92-44b5-9a4e-4199579ab678/volumes" Oct 04 05:52:01 crc kubenswrapper[4770]: I1004 05:52:01.795180 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:52:01 crc kubenswrapper[4770]: I1004 05:52:01.796477 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:52:31 crc kubenswrapper[4770]: I1004 05:52:31.796111 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:52:31 crc kubenswrapper[4770]: I1004 05:52:31.796660 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:53:01 crc kubenswrapper[4770]: I1004 05:53:01.795817 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:53:01 crc kubenswrapper[4770]: I1004 05:53:01.796358 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:53:01 crc kubenswrapper[4770]: I1004 05:53:01.796408 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:53:01 crc kubenswrapper[4770]: I1004 05:53:01.797309 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ef7d94b51b87f586c30b02e12461a333be72c1e193d7af0e3b2c10208c30e361"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:53:01 crc kubenswrapper[4770]: I1004 05:53:01.797373 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://ef7d94b51b87f586c30b02e12461a333be72c1e193d7af0e3b2c10208c30e361" gracePeriod=600 Oct 04 05:53:02 crc kubenswrapper[4770]: I1004 05:53:02.977654 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="ef7d94b51b87f586c30b02e12461a333be72c1e193d7af0e3b2c10208c30e361" exitCode=0 Oct 04 05:53:02 crc kubenswrapper[4770]: I1004 05:53:02.977735 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"ef7d94b51b87f586c30b02e12461a333be72c1e193d7af0e3b2c10208c30e361"} Oct 04 05:53:02 crc kubenswrapper[4770]: I1004 05:53:02.978342 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444"} Oct 04 05:53:02 crc kubenswrapper[4770]: I1004 05:53:02.978367 4770 scope.go:117] "RemoveContainer" containerID="2c4ca8bb667e84a843f0df3599390c057d411c66743260947bd764a2bf8ce1bd" Oct 04 05:53:22 crc kubenswrapper[4770]: I1004 05:53:22.936295 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dnb5s"] Oct 04 05:53:22 crc kubenswrapper[4770]: E1004 05:53:22.937616 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="999af072-bf92-44b5-9a4e-4199579ab678" containerName="registry-server" Oct 04 05:53:22 crc kubenswrapper[4770]: I1004 05:53:22.937636 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="999af072-bf92-44b5-9a4e-4199579ab678" containerName="registry-server" Oct 04 05:53:22 crc kubenswrapper[4770]: E1004 05:53:22.937653 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="999af072-bf92-44b5-9a4e-4199579ab678" containerName="extract-utilities" Oct 04 05:53:22 crc kubenswrapper[4770]: I1004 05:53:22.937662 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="999af072-bf92-44b5-9a4e-4199579ab678" containerName="extract-utilities" Oct 04 05:53:22 crc kubenswrapper[4770]: E1004 05:53:22.937690 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="999af072-bf92-44b5-9a4e-4199579ab678" containerName="extract-content" Oct 04 05:53:22 crc kubenswrapper[4770]: I1004 05:53:22.937697 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="999af072-bf92-44b5-9a4e-4199579ab678" containerName="extract-content" Oct 04 05:53:22 crc kubenswrapper[4770]: I1004 05:53:22.939218 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="999af072-bf92-44b5-9a4e-4199579ab678" containerName="registry-server" Oct 04 05:53:22 crc kubenswrapper[4770]: I1004 05:53:22.950797 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:22 crc kubenswrapper[4770]: I1004 05:53:22.980123 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dnb5s"] Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.120675 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-utilities\") pod \"community-operators-dnb5s\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.120722 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-catalog-content\") pod \"community-operators-dnb5s\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.120802 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n9h4\" (UniqueName: \"kubernetes.io/projected/feb4c770-b40a-423c-80a9-d04e64ca0abf-kube-api-access-8n9h4\") pod \"community-operators-dnb5s\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.222827 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-utilities\") pod \"community-operators-dnb5s\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.222871 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-catalog-content\") pod \"community-operators-dnb5s\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.222960 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n9h4\" (UniqueName: \"kubernetes.io/projected/feb4c770-b40a-423c-80a9-d04e64ca0abf-kube-api-access-8n9h4\") pod \"community-operators-dnb5s\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.223499 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-utilities\") pod \"community-operators-dnb5s\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.223632 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-catalog-content\") pod \"community-operators-dnb5s\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.245079 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n9h4\" (UniqueName: \"kubernetes.io/projected/feb4c770-b40a-423c-80a9-d04e64ca0abf-kube-api-access-8n9h4\") pod \"community-operators-dnb5s\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.276289 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:23 crc kubenswrapper[4770]: I1004 05:53:23.894514 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dnb5s"] Oct 04 05:53:24 crc kubenswrapper[4770]: I1004 05:53:24.197297 4770 generic.go:334] "Generic (PLEG): container finished" podID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerID="1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a" exitCode=0 Oct 04 05:53:24 crc kubenswrapper[4770]: I1004 05:53:24.197401 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnb5s" event={"ID":"feb4c770-b40a-423c-80a9-d04e64ca0abf","Type":"ContainerDied","Data":"1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a"} Oct 04 05:53:24 crc kubenswrapper[4770]: I1004 05:53:24.197435 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnb5s" event={"ID":"feb4c770-b40a-423c-80a9-d04e64ca0abf","Type":"ContainerStarted","Data":"4d85e4ef733d94dbb5234d58322a4d63442e958257a4199582923d1062f48702"} Oct 04 05:53:24 crc kubenswrapper[4770]: E1004 05:53:24.421449 4770 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeb4c770_b40a_423c_80a9_d04e64ca0abf.slice/crio-1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeb4c770_b40a_423c_80a9_d04e64ca0abf.slice/crio-conmon-1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a.scope\": RecentStats: unable to find data in memory cache]" Oct 04 05:53:25 crc kubenswrapper[4770]: I1004 05:53:25.208402 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnb5s" event={"ID":"feb4c770-b40a-423c-80a9-d04e64ca0abf","Type":"ContainerStarted","Data":"8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53"} Oct 04 05:53:26 crc kubenswrapper[4770]: I1004 05:53:26.219877 4770 generic.go:334] "Generic (PLEG): container finished" podID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerID="8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53" exitCode=0 Oct 04 05:53:26 crc kubenswrapper[4770]: I1004 05:53:26.219949 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnb5s" event={"ID":"feb4c770-b40a-423c-80a9-d04e64ca0abf","Type":"ContainerDied","Data":"8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53"} Oct 04 05:53:28 crc kubenswrapper[4770]: I1004 05:53:28.244852 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnb5s" event={"ID":"feb4c770-b40a-423c-80a9-d04e64ca0abf","Type":"ContainerStarted","Data":"e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86"} Oct 04 05:53:28 crc kubenswrapper[4770]: I1004 05:53:28.272665 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dnb5s" podStartSLOduration=3.490282213 podStartE2EDuration="6.272644928s" podCreationTimestamp="2025-10-04 05:53:22 +0000 UTC" firstStartedPulling="2025-10-04 05:53:24.20234618 +0000 UTC m=+10215.494355902" lastFinishedPulling="2025-10-04 05:53:26.984708915 +0000 UTC m=+10218.276718617" observedRunningTime="2025-10-04 05:53:28.269886936 +0000 UTC m=+10219.561896648" watchObservedRunningTime="2025-10-04 05:53:28.272644928 +0000 UTC m=+10219.564654640" Oct 04 05:53:33 crc kubenswrapper[4770]: I1004 05:53:33.277493 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:33 crc kubenswrapper[4770]: I1004 05:53:33.278096 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:33 crc kubenswrapper[4770]: I1004 05:53:33.333118 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:33 crc kubenswrapper[4770]: I1004 05:53:33.407023 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:33 crc kubenswrapper[4770]: I1004 05:53:33.578800 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dnb5s"] Oct 04 05:53:35 crc kubenswrapper[4770]: I1004 05:53:35.338635 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dnb5s" podUID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerName="registry-server" containerID="cri-o://e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86" gracePeriod=2 Oct 04 05:53:35 crc kubenswrapper[4770]: I1004 05:53:35.901729 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:35 crc kubenswrapper[4770]: I1004 05:53:35.954204 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-utilities\") pod \"feb4c770-b40a-423c-80a9-d04e64ca0abf\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " Oct 04 05:53:35 crc kubenswrapper[4770]: I1004 05:53:35.954303 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n9h4\" (UniqueName: \"kubernetes.io/projected/feb4c770-b40a-423c-80a9-d04e64ca0abf-kube-api-access-8n9h4\") pod \"feb4c770-b40a-423c-80a9-d04e64ca0abf\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " Oct 04 05:53:35 crc kubenswrapper[4770]: I1004 05:53:35.954331 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-catalog-content\") pod \"feb4c770-b40a-423c-80a9-d04e64ca0abf\" (UID: \"feb4c770-b40a-423c-80a9-d04e64ca0abf\") " Oct 04 05:53:35 crc kubenswrapper[4770]: I1004 05:53:35.954921 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-utilities" (OuterVolumeSpecName: "utilities") pod "feb4c770-b40a-423c-80a9-d04e64ca0abf" (UID: "feb4c770-b40a-423c-80a9-d04e64ca0abf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:53:35 crc kubenswrapper[4770]: I1004 05:53:35.960902 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feb4c770-b40a-423c-80a9-d04e64ca0abf-kube-api-access-8n9h4" (OuterVolumeSpecName: "kube-api-access-8n9h4") pod "feb4c770-b40a-423c-80a9-d04e64ca0abf" (UID: "feb4c770-b40a-423c-80a9-d04e64ca0abf"). InnerVolumeSpecName "kube-api-access-8n9h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.013612 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "feb4c770-b40a-423c-80a9-d04e64ca0abf" (UID: "feb4c770-b40a-423c-80a9-d04e64ca0abf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.056871 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.056899 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n9h4\" (UniqueName: \"kubernetes.io/projected/feb4c770-b40a-423c-80a9-d04e64ca0abf-kube-api-access-8n9h4\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.056909 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/feb4c770-b40a-423c-80a9-d04e64ca0abf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.379322 4770 generic.go:334] "Generic (PLEG): container finished" podID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerID="e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86" exitCode=0 Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.379367 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnb5s" event={"ID":"feb4c770-b40a-423c-80a9-d04e64ca0abf","Type":"ContainerDied","Data":"e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86"} Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.379405 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnb5s" event={"ID":"feb4c770-b40a-423c-80a9-d04e64ca0abf","Type":"ContainerDied","Data":"4d85e4ef733d94dbb5234d58322a4d63442e958257a4199582923d1062f48702"} Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.379422 4770 scope.go:117] "RemoveContainer" containerID="e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.379550 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnb5s" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.409203 4770 scope.go:117] "RemoveContainer" containerID="8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.430253 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dnb5s"] Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.441237 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dnb5s"] Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.447648 4770 scope.go:117] "RemoveContainer" containerID="1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.500788 4770 scope.go:117] "RemoveContainer" containerID="e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86" Oct 04 05:53:36 crc kubenswrapper[4770]: E1004 05:53:36.501706 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86\": container with ID starting with e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86 not found: ID does not exist" containerID="e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.501752 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86"} err="failed to get container status \"e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86\": rpc error: code = NotFound desc = could not find container \"e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86\": container with ID starting with e8b90b234b9f358a9c9ed50cd7153f7a8b65b8aad8ca206d0380abfbbfb49d86 not found: ID does not exist" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.501779 4770 scope.go:117] "RemoveContainer" containerID="8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53" Oct 04 05:53:36 crc kubenswrapper[4770]: E1004 05:53:36.502059 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53\": container with ID starting with 8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53 not found: ID does not exist" containerID="8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.502098 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53"} err="failed to get container status \"8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53\": rpc error: code = NotFound desc = could not find container \"8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53\": container with ID starting with 8185d651c17ff20246f066018105a5ee977aaba3843942ad08a6ba98eb5bbf53 not found: ID does not exist" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.502124 4770 scope.go:117] "RemoveContainer" containerID="1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a" Oct 04 05:53:36 crc kubenswrapper[4770]: E1004 05:53:36.502360 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a\": container with ID starting with 1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a not found: ID does not exist" containerID="1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a" Oct 04 05:53:36 crc kubenswrapper[4770]: I1004 05:53:36.502384 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a"} err="failed to get container status \"1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a\": rpc error: code = NotFound desc = could not find container \"1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a\": container with ID starting with 1a7c4ae69315cea9fdc550828cc5af33519c018eedce55dd7787bc1d00c0cd0a not found: ID does not exist" Oct 04 05:53:37 crc kubenswrapper[4770]: I1004 05:53:37.394062 4770 generic.go:334] "Generic (PLEG): container finished" podID="9549539a-734a-482a-a374-fbd7be4dc5cc" containerID="e9f3277d28549e2d9783a2fcd1c937d05376654920eed0307d091f77bdd94b86" exitCode=0 Oct 04 05:53:37 crc kubenswrapper[4770]: I1004 05:53:37.394235 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" event={"ID":"9549539a-734a-482a-a374-fbd7be4dc5cc","Type":"ContainerDied","Data":"e9f3277d28549e2d9783a2fcd1c937d05376654920eed0307d091f77bdd94b86"} Oct 04 05:53:37 crc kubenswrapper[4770]: I1004 05:53:37.688277 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feb4c770-b40a-423c-80a9-d04e64ca0abf" path="/var/lib/kubelet/pods/feb4c770-b40a-423c-80a9-d04e64ca0abf/volumes" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.098464 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.123792 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-1\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.123861 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-combined-ca-bundle\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.123942 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-0\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.123978 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-1\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.124091 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ssh-key\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.124146 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-inventory\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.124176 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-0\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.124220 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ceph\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.124250 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-0\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.124284 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-1\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.124325 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws8f6\" (UniqueName: \"kubernetes.io/projected/9549539a-734a-482a-a374-fbd7be4dc5cc-kube-api-access-ws8f6\") pod \"9549539a-734a-482a-a374-fbd7be4dc5cc\" (UID: \"9549539a-734a-482a-a374-fbd7be4dc5cc\") " Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.130088 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9549539a-734a-482a-a374-fbd7be4dc5cc-kube-api-access-ws8f6" (OuterVolumeSpecName: "kube-api-access-ws8f6") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "kube-api-access-ws8f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.133976 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ceph" (OuterVolumeSpecName: "ceph") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.150654 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.166918 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.178184 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.184534 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.235476 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.235530 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.235553 4770 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ceph\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.235570 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.235585 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws8f6\" (UniqueName: \"kubernetes.io/projected/9549539a-734a-482a-a374-fbd7be4dc5cc-kube-api-access-ws8f6\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.235600 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.276417 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-inventory" (OuterVolumeSpecName: "inventory") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.287975 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.288425 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.291992 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.337721 4770 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.337759 4770 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.337770 4770 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.337781 4770 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.338107 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "9549539a-734a-482a-a374-fbd7be4dc5cc" (UID: "9549539a-734a-482a-a374-fbd7be4dc5cc"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.415339 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" event={"ID":"9549539a-734a-482a-a374-fbd7be4dc5cc","Type":"ContainerDied","Data":"42058f536cd0a622d61e30bb028b7b64cacffc512932ce56b144cb09e612c156"} Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.415772 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42058f536cd0a622d61e30bb028b7b64cacffc512932ce56b144cb09e612c156" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.415405 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw" Oct 04 05:53:39 crc kubenswrapper[4770]: I1004 05:53:39.440036 4770 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9549539a-734a-482a-a374-fbd7be4dc5cc-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.163313 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mgr7p"] Oct 04 05:54:20 crc kubenswrapper[4770]: E1004 05:54:20.164912 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9549539a-734a-482a-a374-fbd7be4dc5cc" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.164934 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9549539a-734a-482a-a374-fbd7be4dc5cc" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 04 05:54:20 crc kubenswrapper[4770]: E1004 05:54:20.164976 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerName="extract-content" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.164990 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerName="extract-content" Oct 04 05:54:20 crc kubenswrapper[4770]: E1004 05:54:20.165028 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerName="registry-server" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.165040 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerName="registry-server" Oct 04 05:54:20 crc kubenswrapper[4770]: E1004 05:54:20.165085 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerName="extract-utilities" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.165094 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerName="extract-utilities" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.165355 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="feb4c770-b40a-423c-80a9-d04e64ca0abf" containerName="registry-server" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.165375 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9549539a-734a-482a-a374-fbd7be4dc5cc" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.167655 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.183817 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mgr7p"] Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.264507 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-utilities\") pod \"certified-operators-mgr7p\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.264558 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksnq7\" (UniqueName: \"kubernetes.io/projected/b17cc92a-8d9c-4cce-9303-3feff08fdaee-kube-api-access-ksnq7\") pod \"certified-operators-mgr7p\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.264581 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-catalog-content\") pod \"certified-operators-mgr7p\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.366899 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-utilities\") pod \"certified-operators-mgr7p\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.366975 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksnq7\" (UniqueName: \"kubernetes.io/projected/b17cc92a-8d9c-4cce-9303-3feff08fdaee-kube-api-access-ksnq7\") pod \"certified-operators-mgr7p\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.367035 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-catalog-content\") pod \"certified-operators-mgr7p\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.367382 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-utilities\") pod \"certified-operators-mgr7p\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.367546 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-catalog-content\") pod \"certified-operators-mgr7p\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.391854 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksnq7\" (UniqueName: \"kubernetes.io/projected/b17cc92a-8d9c-4cce-9303-3feff08fdaee-kube-api-access-ksnq7\") pod \"certified-operators-mgr7p\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:20 crc kubenswrapper[4770]: I1004 05:54:20.507401 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:21 crc kubenswrapper[4770]: I1004 05:54:21.015864 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mgr7p"] Oct 04 05:54:21 crc kubenswrapper[4770]: I1004 05:54:21.840800 4770 generic.go:334] "Generic (PLEG): container finished" podID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerID="5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa" exitCode=0 Oct 04 05:54:21 crc kubenswrapper[4770]: I1004 05:54:21.841158 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgr7p" event={"ID":"b17cc92a-8d9c-4cce-9303-3feff08fdaee","Type":"ContainerDied","Data":"5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa"} Oct 04 05:54:21 crc kubenswrapper[4770]: I1004 05:54:21.841183 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgr7p" event={"ID":"b17cc92a-8d9c-4cce-9303-3feff08fdaee","Type":"ContainerStarted","Data":"dbb3c131648794ebcd573a97ba1c51629d0d3c98bb8e220274e5cbdb85d5afc5"} Oct 04 05:54:23 crc kubenswrapper[4770]: I1004 05:54:23.866124 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgr7p" event={"ID":"b17cc92a-8d9c-4cce-9303-3feff08fdaee","Type":"ContainerStarted","Data":"9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2"} Oct 04 05:54:26 crc kubenswrapper[4770]: E1004 05:54:26.316906 4770 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.68:54670->38.102.83.68:36801: write tcp 38.102.83.68:54670->38.102.83.68:36801: write: broken pipe Oct 04 05:54:27 crc kubenswrapper[4770]: I1004 05:54:27.917446 4770 generic.go:334] "Generic (PLEG): container finished" podID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerID="9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2" exitCode=0 Oct 04 05:54:27 crc kubenswrapper[4770]: I1004 05:54:27.917517 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgr7p" event={"ID":"b17cc92a-8d9c-4cce-9303-3feff08fdaee","Type":"ContainerDied","Data":"9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2"} Oct 04 05:54:32 crc kubenswrapper[4770]: I1004 05:54:32.975517 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgr7p" event={"ID":"b17cc92a-8d9c-4cce-9303-3feff08fdaee","Type":"ContainerStarted","Data":"519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4"} Oct 04 05:54:33 crc kubenswrapper[4770]: I1004 05:54:33.012812 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mgr7p" podStartSLOduration=2.4469832 podStartE2EDuration="13.012786879s" podCreationTimestamp="2025-10-04 05:54:20 +0000 UTC" firstStartedPulling="2025-10-04 05:54:21.844886281 +0000 UTC m=+10273.136896003" lastFinishedPulling="2025-10-04 05:54:32.41068996 +0000 UTC m=+10283.702699682" observedRunningTime="2025-10-04 05:54:33.001838943 +0000 UTC m=+10284.293848655" watchObservedRunningTime="2025-10-04 05:54:33.012786879 +0000 UTC m=+10284.304796601" Oct 04 05:54:40 crc kubenswrapper[4770]: I1004 05:54:40.508394 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:40 crc kubenswrapper[4770]: I1004 05:54:40.509201 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:42 crc kubenswrapper[4770]: I1004 05:54:42.092097 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mgr7p" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerName="registry-server" probeResult="failure" output=< Oct 04 05:54:42 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 05:54:42 crc kubenswrapper[4770]: > Oct 04 05:54:50 crc kubenswrapper[4770]: I1004 05:54:50.559705 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:50 crc kubenswrapper[4770]: I1004 05:54:50.611825 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:51 crc kubenswrapper[4770]: I1004 05:54:51.366470 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mgr7p"] Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.157399 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mgr7p" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerName="registry-server" containerID="cri-o://519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4" gracePeriod=2 Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.687536 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.854786 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-catalog-content\") pod \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.855273 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-utilities\") pod \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.855504 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksnq7\" (UniqueName: \"kubernetes.io/projected/b17cc92a-8d9c-4cce-9303-3feff08fdaee-kube-api-access-ksnq7\") pod \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\" (UID: \"b17cc92a-8d9c-4cce-9303-3feff08fdaee\") " Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.859637 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-utilities" (OuterVolumeSpecName: "utilities") pod "b17cc92a-8d9c-4cce-9303-3feff08fdaee" (UID: "b17cc92a-8d9c-4cce-9303-3feff08fdaee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.864521 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b17cc92a-8d9c-4cce-9303-3feff08fdaee-kube-api-access-ksnq7" (OuterVolumeSpecName: "kube-api-access-ksnq7") pod "b17cc92a-8d9c-4cce-9303-3feff08fdaee" (UID: "b17cc92a-8d9c-4cce-9303-3feff08fdaee"). InnerVolumeSpecName "kube-api-access-ksnq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.909357 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b17cc92a-8d9c-4cce-9303-3feff08fdaee" (UID: "b17cc92a-8d9c-4cce-9303-3feff08fdaee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.957839 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.957868 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksnq7\" (UniqueName: \"kubernetes.io/projected/b17cc92a-8d9c-4cce-9303-3feff08fdaee-kube-api-access-ksnq7\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:52 crc kubenswrapper[4770]: I1004 05:54:52.957877 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b17cc92a-8d9c-4cce-9303-3feff08fdaee-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.168731 4770 generic.go:334] "Generic (PLEG): container finished" podID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerID="519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4" exitCode=0 Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.168779 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgr7p" event={"ID":"b17cc92a-8d9c-4cce-9303-3feff08fdaee","Type":"ContainerDied","Data":"519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4"} Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.168809 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgr7p" event={"ID":"b17cc92a-8d9c-4cce-9303-3feff08fdaee","Type":"ContainerDied","Data":"dbb3c131648794ebcd573a97ba1c51629d0d3c98bb8e220274e5cbdb85d5afc5"} Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.168826 4770 scope.go:117] "RemoveContainer" containerID="519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.170118 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mgr7p" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.192292 4770 scope.go:117] "RemoveContainer" containerID="9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.210030 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mgr7p"] Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.219174 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mgr7p"] Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.233602 4770 scope.go:117] "RemoveContainer" containerID="5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.298561 4770 scope.go:117] "RemoveContainer" containerID="519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4" Oct 04 05:54:53 crc kubenswrapper[4770]: E1004 05:54:53.299638 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4\": container with ID starting with 519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4 not found: ID does not exist" containerID="519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.299901 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4"} err="failed to get container status \"519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4\": rpc error: code = NotFound desc = could not find container \"519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4\": container with ID starting with 519e6764457f410a2c58925b36f3d40611a00b3b292241097b32153e56661fb4 not found: ID does not exist" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.299931 4770 scope.go:117] "RemoveContainer" containerID="9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2" Oct 04 05:54:53 crc kubenswrapper[4770]: E1004 05:54:53.300452 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2\": container with ID starting with 9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2 not found: ID does not exist" containerID="9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.300489 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2"} err="failed to get container status \"9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2\": rpc error: code = NotFound desc = could not find container \"9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2\": container with ID starting with 9ad541e9c1eaef1da427d03283c0c5630d2f5c374acd33dc90b6c339c58f6ef2 not found: ID does not exist" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.300509 4770 scope.go:117] "RemoveContainer" containerID="5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa" Oct 04 05:54:53 crc kubenswrapper[4770]: E1004 05:54:53.303166 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa\": container with ID starting with 5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa not found: ID does not exist" containerID="5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.303208 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa"} err="failed to get container status \"5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa\": rpc error: code = NotFound desc = could not find container \"5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa\": container with ID starting with 5f1cc6770caba0a5d145ba16bfa5f794a730d9e6356df056e4b7bd947c8dc7aa not found: ID does not exist" Oct 04 05:54:53 crc kubenswrapper[4770]: I1004 05:54:53.709992 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" path="/var/lib/kubelet/pods/b17cc92a-8d9c-4cce-9303-3feff08fdaee/volumes" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.089190 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m5l6b"] Oct 04 05:55:26 crc kubenswrapper[4770]: E1004 05:55:26.090416 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerName="extract-content" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.090438 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerName="extract-content" Oct 04 05:55:26 crc kubenswrapper[4770]: E1004 05:55:26.090506 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerName="extract-utilities" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.090515 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerName="extract-utilities" Oct 04 05:55:26 crc kubenswrapper[4770]: E1004 05:55:26.090528 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerName="registry-server" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.090537 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerName="registry-server" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.090818 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b17cc92a-8d9c-4cce-9303-3feff08fdaee" containerName="registry-server" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.092834 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.099594 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m5l6b"] Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.196332 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-catalog-content\") pod \"redhat-operators-m5l6b\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.196456 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5bxj\" (UniqueName: \"kubernetes.io/projected/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-kube-api-access-t5bxj\") pod \"redhat-operators-m5l6b\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.196612 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-utilities\") pod \"redhat-operators-m5l6b\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.298298 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5bxj\" (UniqueName: \"kubernetes.io/projected/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-kube-api-access-t5bxj\") pod \"redhat-operators-m5l6b\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.298446 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-utilities\") pod \"redhat-operators-m5l6b\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.298562 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-catalog-content\") pod \"redhat-operators-m5l6b\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.299183 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-catalog-content\") pod \"redhat-operators-m5l6b\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.299273 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-utilities\") pod \"redhat-operators-m5l6b\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.319445 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5bxj\" (UniqueName: \"kubernetes.io/projected/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-kube-api-access-t5bxj\") pod \"redhat-operators-m5l6b\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:26 crc kubenswrapper[4770]: I1004 05:55:26.431867 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:27 crc kubenswrapper[4770]: I1004 05:55:27.408172 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m5l6b"] Oct 04 05:55:27 crc kubenswrapper[4770]: I1004 05:55:27.534419 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5l6b" event={"ID":"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7","Type":"ContainerStarted","Data":"1cbcd95ab076ef77af8345b84ce5b02c684ddbe922ff4b570e9ee25be09e38e5"} Oct 04 05:55:28 crc kubenswrapper[4770]: I1004 05:55:28.547795 4770 generic.go:334] "Generic (PLEG): container finished" podID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerID="642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce" exitCode=0 Oct 04 05:55:28 crc kubenswrapper[4770]: I1004 05:55:28.547906 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5l6b" event={"ID":"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7","Type":"ContainerDied","Data":"642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce"} Oct 04 05:55:30 crc kubenswrapper[4770]: I1004 05:55:30.569236 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5l6b" event={"ID":"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7","Type":"ContainerStarted","Data":"2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7"} Oct 04 05:55:31 crc kubenswrapper[4770]: I1004 05:55:31.795418 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:55:31 crc kubenswrapper[4770]: I1004 05:55:31.795632 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:55:32 crc kubenswrapper[4770]: I1004 05:55:32.607930 4770 generic.go:334] "Generic (PLEG): container finished" podID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerID="2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7" exitCode=0 Oct 04 05:55:32 crc kubenswrapper[4770]: I1004 05:55:32.608267 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5l6b" event={"ID":"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7","Type":"ContainerDied","Data":"2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7"} Oct 04 05:55:36 crc kubenswrapper[4770]: E1004 05:55:36.246373 4770 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.68:55132->38.102.83.68:36801: write tcp 38.102.83.68:55132->38.102.83.68:36801: write: broken pipe Oct 04 05:55:38 crc kubenswrapper[4770]: I1004 05:55:38.680891 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5l6b" event={"ID":"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7","Type":"ContainerStarted","Data":"0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1"} Oct 04 05:55:40 crc kubenswrapper[4770]: I1004 05:55:40.724636 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m5l6b" podStartSLOduration=7.277295413 podStartE2EDuration="14.724613842s" podCreationTimestamp="2025-10-04 05:55:26 +0000 UTC" firstStartedPulling="2025-10-04 05:55:28.549981412 +0000 UTC m=+10339.841991124" lastFinishedPulling="2025-10-04 05:55:35.997299841 +0000 UTC m=+10347.289309553" observedRunningTime="2025-10-04 05:55:40.720255317 +0000 UTC m=+10352.012265029" watchObservedRunningTime="2025-10-04 05:55:40.724613842 +0000 UTC m=+10352.016623544" Oct 04 05:55:46 crc kubenswrapper[4770]: I1004 05:55:46.432208 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:46 crc kubenswrapper[4770]: I1004 05:55:46.433626 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:46 crc kubenswrapper[4770]: I1004 05:55:46.482714 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:46 crc kubenswrapper[4770]: I1004 05:55:46.823126 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:48 crc kubenswrapper[4770]: I1004 05:55:48.804354 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m5l6b"] Oct 04 05:55:49 crc kubenswrapper[4770]: I1004 05:55:49.809562 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m5l6b" podUID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerName="registry-server" containerID="cri-o://0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1" gracePeriod=2 Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.352492 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.520359 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-catalog-content\") pod \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.520458 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5bxj\" (UniqueName: \"kubernetes.io/projected/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-kube-api-access-t5bxj\") pod \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.520555 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-utilities\") pod \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\" (UID: \"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7\") " Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.521563 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-utilities" (OuterVolumeSpecName: "utilities") pod "9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" (UID: "9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.608291 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" (UID: "9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.622993 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.623027 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.820533 4770 generic.go:334] "Generic (PLEG): container finished" podID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerID="0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1" exitCode=0 Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.820594 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5l6b" event={"ID":"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7","Type":"ContainerDied","Data":"0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1"} Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.820608 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m5l6b" Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.820836 4770 scope.go:117] "RemoveContainer" containerID="0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1" Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.820825 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m5l6b" event={"ID":"9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7","Type":"ContainerDied","Data":"1cbcd95ab076ef77af8345b84ce5b02c684ddbe922ff4b570e9ee25be09e38e5"} Oct 04 05:55:50 crc kubenswrapper[4770]: I1004 05:55:50.842667 4770 scope.go:117] "RemoveContainer" containerID="2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.157045 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-kube-api-access-t5bxj" (OuterVolumeSpecName: "kube-api-access-t5bxj") pod "9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" (UID: "9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7"). InnerVolumeSpecName "kube-api-access-t5bxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.166390 4770 scope.go:117] "RemoveContainer" containerID="642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.245702 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5bxj\" (UniqueName: \"kubernetes.io/projected/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7-kube-api-access-t5bxj\") on node \"crc\" DevicePath \"\"" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.366021 4770 scope.go:117] "RemoveContainer" containerID="0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1" Oct 04 05:55:51 crc kubenswrapper[4770]: E1004 05:55:51.366905 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1\": container with ID starting with 0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1 not found: ID does not exist" containerID="0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.366937 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1"} err="failed to get container status \"0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1\": rpc error: code = NotFound desc = could not find container \"0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1\": container with ID starting with 0ca0c63c55f397c874cd6b9b7beede11725d505d6289dda504c88b074adad5b1 not found: ID does not exist" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.366958 4770 scope.go:117] "RemoveContainer" containerID="2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7" Oct 04 05:55:51 crc kubenswrapper[4770]: E1004 05:55:51.367431 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7\": container with ID starting with 2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7 not found: ID does not exist" containerID="2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.367459 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7"} err="failed to get container status \"2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7\": rpc error: code = NotFound desc = could not find container \"2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7\": container with ID starting with 2f7271d48af2b8e32e6b97032fffaad12df78e231a01f9874920ef01a7593ce7 not found: ID does not exist" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.367473 4770 scope.go:117] "RemoveContainer" containerID="642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce" Oct 04 05:55:51 crc kubenswrapper[4770]: E1004 05:55:51.368157 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce\": container with ID starting with 642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce not found: ID does not exist" containerID="642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.368176 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce"} err="failed to get container status \"642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce\": rpc error: code = NotFound desc = could not find container \"642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce\": container with ID starting with 642ffe3018be3ed3ab0e961448ba5390c09fc0a7fab14dbf8ab5dd74abaa66ce not found: ID does not exist" Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.459822 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m5l6b"] Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.468768 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m5l6b"] Oct 04 05:55:51 crc kubenswrapper[4770]: I1004 05:55:51.686058 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" path="/var/lib/kubelet/pods/9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7/volumes" Oct 04 05:56:01 crc kubenswrapper[4770]: I1004 05:56:01.796055 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:56:01 crc kubenswrapper[4770]: I1004 05:56:01.796416 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:56:06 crc kubenswrapper[4770]: I1004 05:56:06.789328 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 04 05:56:06 crc kubenswrapper[4770]: I1004 05:56:06.791285 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="dcea39e7-a0fb-4831-81de-7cbecec5c963" containerName="adoption" containerID="cri-o://f4191c4e27e4d21d7679fd1172e3ac6c5627f44f8a756a032d06c4cf5c762683" gracePeriod=30 Oct 04 05:56:31 crc kubenswrapper[4770]: I1004 05:56:31.796516 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:56:31 crc kubenswrapper[4770]: I1004 05:56:31.797144 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:56:31 crc kubenswrapper[4770]: I1004 05:56:31.797193 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 05:56:31 crc kubenswrapper[4770]: I1004 05:56:31.797981 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:56:31 crc kubenswrapper[4770]: I1004 05:56:31.798053 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" gracePeriod=600 Oct 04 05:56:31 crc kubenswrapper[4770]: E1004 05:56:31.970880 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:56:32 crc kubenswrapper[4770]: I1004 05:56:32.263254 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" exitCode=0 Oct 04 05:56:32 crc kubenswrapper[4770]: I1004 05:56:32.263835 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444"} Oct 04 05:56:32 crc kubenswrapper[4770]: I1004 05:56:32.264130 4770 scope.go:117] "RemoveContainer" containerID="ef7d94b51b87f586c30b02e12461a333be72c1e193d7af0e3b2c10208c30e361" Oct 04 05:56:32 crc kubenswrapper[4770]: I1004 05:56:32.268112 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:56:32 crc kubenswrapper[4770]: E1004 05:56:32.268623 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:56:37 crc kubenswrapper[4770]: I1004 05:56:37.320040 4770 generic.go:334] "Generic (PLEG): container finished" podID="dcea39e7-a0fb-4831-81de-7cbecec5c963" containerID="f4191c4e27e4d21d7679fd1172e3ac6c5627f44f8a756a032d06c4cf5c762683" exitCode=137 Oct 04 05:56:37 crc kubenswrapper[4770]: I1004 05:56:37.320107 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"dcea39e7-a0fb-4831-81de-7cbecec5c963","Type":"ContainerDied","Data":"f4191c4e27e4d21d7679fd1172e3ac6c5627f44f8a756a032d06c4cf5c762683"} Oct 04 05:56:37 crc kubenswrapper[4770]: I1004 05:56:37.461442 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 04 05:56:37 crc kubenswrapper[4770]: I1004 05:56:37.580269 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgzcg\" (UniqueName: \"kubernetes.io/projected/dcea39e7-a0fb-4831-81de-7cbecec5c963-kube-api-access-sgzcg\") pod \"dcea39e7-a0fb-4831-81de-7cbecec5c963\" (UID: \"dcea39e7-a0fb-4831-81de-7cbecec5c963\") " Oct 04 05:56:37 crc kubenswrapper[4770]: I1004 05:56:37.580986 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83d8cec1-7494-446d-a411-cea543767bc5\") pod \"dcea39e7-a0fb-4831-81de-7cbecec5c963\" (UID: \"dcea39e7-a0fb-4831-81de-7cbecec5c963\") " Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.044297 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcea39e7-a0fb-4831-81de-7cbecec5c963-kube-api-access-sgzcg" (OuterVolumeSpecName: "kube-api-access-sgzcg") pod "dcea39e7-a0fb-4831-81de-7cbecec5c963" (UID: "dcea39e7-a0fb-4831-81de-7cbecec5c963"). InnerVolumeSpecName "kube-api-access-sgzcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.087211 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83d8cec1-7494-446d-a411-cea543767bc5" (OuterVolumeSpecName: "mariadb-data") pod "dcea39e7-a0fb-4831-81de-7cbecec5c963" (UID: "dcea39e7-a0fb-4831-81de-7cbecec5c963"). InnerVolumeSpecName "pvc-83d8cec1-7494-446d-a411-cea543767bc5". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.092260 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgzcg\" (UniqueName: \"kubernetes.io/projected/dcea39e7-a0fb-4831-81de-7cbecec5c963-kube-api-access-sgzcg\") on node \"crc\" DevicePath \"\"" Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.092324 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-83d8cec1-7494-446d-a411-cea543767bc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83d8cec1-7494-446d-a411-cea543767bc5\") on node \"crc\" " Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.126737 4770 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.126929 4770 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-83d8cec1-7494-446d-a411-cea543767bc5" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83d8cec1-7494-446d-a411-cea543767bc5") on node "crc" Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.194647 4770 reconciler_common.go:293] "Volume detached for volume \"pvc-83d8cec1-7494-446d-a411-cea543767bc5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83d8cec1-7494-446d-a411-cea543767bc5\") on node \"crc\" DevicePath \"\"" Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.332188 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"dcea39e7-a0fb-4831-81de-7cbecec5c963","Type":"ContainerDied","Data":"99a1ff1e2ad3c16d33bc882d2fc06fca73aac623f8d1867b26a5a7165c776ea7"} Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.332242 4770 scope.go:117] "RemoveContainer" containerID="f4191c4e27e4d21d7679fd1172e3ac6c5627f44f8a756a032d06c4cf5c762683" Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.332261 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.366413 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 04 05:56:38 crc kubenswrapper[4770]: I1004 05:56:38.374998 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Oct 04 05:56:39 crc kubenswrapper[4770]: I1004 05:56:39.046480 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 04 05:56:39 crc kubenswrapper[4770]: I1004 05:56:39.047076 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="b7daf327-2122-4ef8-8352-c41822f77200" containerName="adoption" containerID="cri-o://3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec" gracePeriod=30 Oct 04 05:56:39 crc kubenswrapper[4770]: I1004 05:56:39.686315 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcea39e7-a0fb-4831-81de-7cbecec5c963" path="/var/lib/kubelet/pods/dcea39e7-a0fb-4831-81de-7cbecec5c963/volumes" Oct 04 05:56:43 crc kubenswrapper[4770]: I1004 05:56:43.674043 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:56:43 crc kubenswrapper[4770]: E1004 05:56:43.674841 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:56:57 crc kubenswrapper[4770]: I1004 05:56:57.676128 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:56:57 crc kubenswrapper[4770]: E1004 05:56:57.676998 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.651138 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.665199 4770 generic.go:334] "Generic (PLEG): container finished" podID="b7daf327-2122-4ef8-8352-c41822f77200" containerID="3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec" exitCode=137 Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.665247 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"b7daf327-2122-4ef8-8352-c41822f77200","Type":"ContainerDied","Data":"3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec"} Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.665276 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"b7daf327-2122-4ef8-8352-c41822f77200","Type":"ContainerDied","Data":"3aff780f573332b464e041c318c5d5a034e4308f77f985ca41979a55542f25f3"} Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.665292 4770 scope.go:117] "RemoveContainer" containerID="3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.665409 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.682461 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/b7daf327-2122-4ef8-8352-c41822f77200-ovn-data-cert\") pod \"b7daf327-2122-4ef8-8352-c41822f77200\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.683323 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\") pod \"b7daf327-2122-4ef8-8352-c41822f77200\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.683368 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk4nc\" (UniqueName: \"kubernetes.io/projected/b7daf327-2122-4ef8-8352-c41822f77200-kube-api-access-pk4nc\") pod \"b7daf327-2122-4ef8-8352-c41822f77200\" (UID: \"b7daf327-2122-4ef8-8352-c41822f77200\") " Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.692784 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7daf327-2122-4ef8-8352-c41822f77200-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "b7daf327-2122-4ef8-8352-c41822f77200" (UID: "b7daf327-2122-4ef8-8352-c41822f77200"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.702198 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7daf327-2122-4ef8-8352-c41822f77200-kube-api-access-pk4nc" (OuterVolumeSpecName: "kube-api-access-pk4nc") pod "b7daf327-2122-4ef8-8352-c41822f77200" (UID: "b7daf327-2122-4ef8-8352-c41822f77200"). InnerVolumeSpecName "kube-api-access-pk4nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.712236 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b" (OuterVolumeSpecName: "ovn-data") pod "b7daf327-2122-4ef8-8352-c41822f77200" (UID: "b7daf327-2122-4ef8-8352-c41822f77200"). InnerVolumeSpecName "pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.745930 4770 scope.go:117] "RemoveContainer" containerID="3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec" Oct 04 05:57:09 crc kubenswrapper[4770]: E1004 05:57:09.746452 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec\": container with ID starting with 3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec not found: ID does not exist" containerID="3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.746492 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec"} err="failed to get container status \"3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec\": rpc error: code = NotFound desc = could not find container \"3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec\": container with ID starting with 3558a9e15e67b32c21db4452ea90a275dd9e68e7c4bf2de4b624e1e1ad4697ec not found: ID does not exist" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.787755 4770 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/b7daf327-2122-4ef8-8352-c41822f77200-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.787792 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk4nc\" (UniqueName: \"kubernetes.io/projected/b7daf327-2122-4ef8-8352-c41822f77200-kube-api-access-pk4nc\") on node \"crc\" DevicePath \"\"" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.787818 4770 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\") on node \"crc\" " Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.816157 4770 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.816299 4770 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b") on node "crc" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.889895 4770 reconciler_common.go:293] "Volume detached for volume \"pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94b3112b-77e2-4441-9a0e-928a560d5c7b\") on node \"crc\" DevicePath \"\"" Oct 04 05:57:09 crc kubenswrapper[4770]: I1004 05:57:09.996348 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 04 05:57:10 crc kubenswrapper[4770]: I1004 05:57:10.007901 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Oct 04 05:57:11 crc kubenswrapper[4770]: I1004 05:57:11.686533 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7daf327-2122-4ef8-8352-c41822f77200" path="/var/lib/kubelet/pods/b7daf327-2122-4ef8-8352-c41822f77200/volumes" Oct 04 05:57:12 crc kubenswrapper[4770]: I1004 05:57:12.673802 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:57:12 crc kubenswrapper[4770]: E1004 05:57:12.674429 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:57:23 crc kubenswrapper[4770]: I1004 05:57:23.674747 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:57:23 crc kubenswrapper[4770]: E1004 05:57:23.676039 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:57:38 crc kubenswrapper[4770]: I1004 05:57:38.674241 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:57:38 crc kubenswrapper[4770]: E1004 05:57:38.676332 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:57:53 crc kubenswrapper[4770]: I1004 05:57:53.674477 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:57:53 crc kubenswrapper[4770]: E1004 05:57:53.675130 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:58:08 crc kubenswrapper[4770]: I1004 05:58:08.674852 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:58:08 crc kubenswrapper[4770]: E1004 05:58:08.675782 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.204642 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c885t/must-gather-tnff4"] Oct 04 05:58:21 crc kubenswrapper[4770]: E1004 05:58:21.205706 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerName="extract-content" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.205722 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerName="extract-content" Oct 04 05:58:21 crc kubenswrapper[4770]: E1004 05:58:21.205744 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerName="registry-server" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.205754 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerName="registry-server" Oct 04 05:58:21 crc kubenswrapper[4770]: E1004 05:58:21.205777 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcea39e7-a0fb-4831-81de-7cbecec5c963" containerName="adoption" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.205784 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcea39e7-a0fb-4831-81de-7cbecec5c963" containerName="adoption" Oct 04 05:58:21 crc kubenswrapper[4770]: E1004 05:58:21.205801 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerName="extract-utilities" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.205810 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerName="extract-utilities" Oct 04 05:58:21 crc kubenswrapper[4770]: E1004 05:58:21.205830 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7daf327-2122-4ef8-8352-c41822f77200" containerName="adoption" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.205838 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7daf327-2122-4ef8-8352-c41822f77200" containerName="adoption" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.206118 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcea39e7-a0fb-4831-81de-7cbecec5c963" containerName="adoption" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.206166 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7daf327-2122-4ef8-8352-c41822f77200" containerName="adoption" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.206197 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c02e0c9-0b2b-4f92-b461-92dc65b1e2b7" containerName="registry-server" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.213227 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.218544 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-c885t"/"openshift-service-ca.crt" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.218840 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-c885t"/"kube-root-ca.crt" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.219002 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-c885t"/"default-dockercfg-t88zf" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.234666 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-c885t/must-gather-tnff4"] Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.259304 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsfp6\" (UniqueName: \"kubernetes.io/projected/b0ad49de-f48e-451b-9382-e0729a649190-kube-api-access-gsfp6\") pod \"must-gather-tnff4\" (UID: \"b0ad49de-f48e-451b-9382-e0729a649190\") " pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.259508 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0ad49de-f48e-451b-9382-e0729a649190-must-gather-output\") pod \"must-gather-tnff4\" (UID: \"b0ad49de-f48e-451b-9382-e0729a649190\") " pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.363315 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsfp6\" (UniqueName: \"kubernetes.io/projected/b0ad49de-f48e-451b-9382-e0729a649190-kube-api-access-gsfp6\") pod \"must-gather-tnff4\" (UID: \"b0ad49de-f48e-451b-9382-e0729a649190\") " pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.363556 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0ad49de-f48e-451b-9382-e0729a649190-must-gather-output\") pod \"must-gather-tnff4\" (UID: \"b0ad49de-f48e-451b-9382-e0729a649190\") " pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.364366 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0ad49de-f48e-451b-9382-e0729a649190-must-gather-output\") pod \"must-gather-tnff4\" (UID: \"b0ad49de-f48e-451b-9382-e0729a649190\") " pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.387726 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsfp6\" (UniqueName: \"kubernetes.io/projected/b0ad49de-f48e-451b-9382-e0729a649190-kube-api-access-gsfp6\") pod \"must-gather-tnff4\" (UID: \"b0ad49de-f48e-451b-9382-e0729a649190\") " pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.554619 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 05:58:21 crc kubenswrapper[4770]: I1004 05:58:21.679967 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:58:21 crc kubenswrapper[4770]: E1004 05:58:21.682712 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:58:22 crc kubenswrapper[4770]: I1004 05:58:22.151925 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-c885t/must-gather-tnff4"] Oct 04 05:58:22 crc kubenswrapper[4770]: I1004 05:58:22.157791 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:58:22 crc kubenswrapper[4770]: I1004 05:58:22.564321 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/must-gather-tnff4" event={"ID":"b0ad49de-f48e-451b-9382-e0729a649190","Type":"ContainerStarted","Data":"a15436a03ba8d3551f5b5086836c9c6231c87685f0fa15e7abb531b62dab13d4"} Oct 04 05:58:29 crc kubenswrapper[4770]: I1004 05:58:29.648397 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/must-gather-tnff4" event={"ID":"b0ad49de-f48e-451b-9382-e0729a649190","Type":"ContainerStarted","Data":"73f14e7309b5309b7c2490eb028a1eea77e7a587862c7c1a67048ec344d0b3cf"} Oct 04 05:58:30 crc kubenswrapper[4770]: I1004 05:58:30.662606 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/must-gather-tnff4" event={"ID":"b0ad49de-f48e-451b-9382-e0729a649190","Type":"ContainerStarted","Data":"afe7316e47faa286fde3ea0d7559f989e5fff996be69c1f89695e896cf2b4509"} Oct 04 05:58:33 crc kubenswrapper[4770]: E1004 05:58:33.130298 4770 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.68:41370->38.102.83.68:36801: write tcp 38.102.83.68:41370->38.102.83.68:36801: write: broken pipe Oct 04 05:58:34 crc kubenswrapper[4770]: I1004 05:58:34.839160 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c885t/must-gather-tnff4" podStartSLOduration=7.579477604 podStartE2EDuration="13.839137831s" podCreationTimestamp="2025-10-04 05:58:21 +0000 UTC" firstStartedPulling="2025-10-04 05:58:22.157597114 +0000 UTC m=+10513.449606826" lastFinishedPulling="2025-10-04 05:58:28.417257341 +0000 UTC m=+10519.709267053" observedRunningTime="2025-10-04 05:58:30.686452646 +0000 UTC m=+10521.978462358" watchObservedRunningTime="2025-10-04 05:58:34.839137831 +0000 UTC m=+10526.131147543" Oct 04 05:58:34 crc kubenswrapper[4770]: I1004 05:58:34.845973 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c885t/crc-debug-zh67x"] Oct 04 05:58:34 crc kubenswrapper[4770]: I1004 05:58:34.847509 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 05:58:34 crc kubenswrapper[4770]: I1004 05:58:34.908144 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06582714-b3e5-4206-9a81-ecb43ec9513d-host\") pod \"crc-debug-zh67x\" (UID: \"06582714-b3e5-4206-9a81-ecb43ec9513d\") " pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 05:58:34 crc kubenswrapper[4770]: I1004 05:58:34.908202 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lgrr\" (UniqueName: \"kubernetes.io/projected/06582714-b3e5-4206-9a81-ecb43ec9513d-kube-api-access-4lgrr\") pod \"crc-debug-zh67x\" (UID: \"06582714-b3e5-4206-9a81-ecb43ec9513d\") " pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 05:58:35 crc kubenswrapper[4770]: I1004 05:58:35.010988 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06582714-b3e5-4206-9a81-ecb43ec9513d-host\") pod \"crc-debug-zh67x\" (UID: \"06582714-b3e5-4206-9a81-ecb43ec9513d\") " pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 05:58:35 crc kubenswrapper[4770]: I1004 05:58:35.011063 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lgrr\" (UniqueName: \"kubernetes.io/projected/06582714-b3e5-4206-9a81-ecb43ec9513d-kube-api-access-4lgrr\") pod \"crc-debug-zh67x\" (UID: \"06582714-b3e5-4206-9a81-ecb43ec9513d\") " pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 05:58:35 crc kubenswrapper[4770]: I1004 05:58:35.011420 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06582714-b3e5-4206-9a81-ecb43ec9513d-host\") pod \"crc-debug-zh67x\" (UID: \"06582714-b3e5-4206-9a81-ecb43ec9513d\") " pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 05:58:35 crc kubenswrapper[4770]: I1004 05:58:35.037674 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lgrr\" (UniqueName: \"kubernetes.io/projected/06582714-b3e5-4206-9a81-ecb43ec9513d-kube-api-access-4lgrr\") pod \"crc-debug-zh67x\" (UID: \"06582714-b3e5-4206-9a81-ecb43ec9513d\") " pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 05:58:35 crc kubenswrapper[4770]: I1004 05:58:35.167835 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 05:58:35 crc kubenswrapper[4770]: I1004 05:58:35.673528 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:58:35 crc kubenswrapper[4770]: E1004 05:58:35.674377 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:58:35 crc kubenswrapper[4770]: I1004 05:58:35.729060 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/crc-debug-zh67x" event={"ID":"06582714-b3e5-4206-9a81-ecb43ec9513d","Type":"ContainerStarted","Data":"9f70f3f6154bd59c9cc975ec529c9c849aee681ca03d8c1b66b7c77df65ccb43"} Oct 04 05:58:48 crc kubenswrapper[4770]: I1004 05:58:48.674566 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:58:48 crc kubenswrapper[4770]: E1004 05:58:48.675493 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:59:02 crc kubenswrapper[4770]: I1004 05:59:02.674503 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:59:02 crc kubenswrapper[4770]: E1004 05:59:02.675276 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:59:03 crc kubenswrapper[4770]: E1004 05:59:03.200859 4770 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Oct 04 05:59:03 crc kubenswrapper[4770]: E1004 05:59:03.201379 4770 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4lgrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-zh67x_openshift-must-gather-c885t(06582714-b3e5-4206-9a81-ecb43ec9513d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 05:59:03 crc kubenswrapper[4770]: E1004 05:59:03.202722 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-c885t/crc-debug-zh67x" podUID="06582714-b3e5-4206-9a81-ecb43ec9513d" Oct 04 05:59:04 crc kubenswrapper[4770]: E1004 05:59:04.009559 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-c885t/crc-debug-zh67x" podUID="06582714-b3e5-4206-9a81-ecb43ec9513d" Oct 04 05:59:15 crc kubenswrapper[4770]: I1004 05:59:15.673629 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:59:15 crc kubenswrapper[4770]: E1004 05:59:15.674359 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:59:19 crc kubenswrapper[4770]: I1004 05:59:19.151432 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/crc-debug-zh67x" event={"ID":"06582714-b3e5-4206-9a81-ecb43ec9513d","Type":"ContainerStarted","Data":"f2384b679082368582479583a5c36a28ab4d7a321803c54c10c1fdb1166afd84"} Oct 04 05:59:19 crc kubenswrapper[4770]: I1004 05:59:19.183729 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c885t/crc-debug-zh67x" podStartSLOduration=2.288679336 podStartE2EDuration="45.183708769s" podCreationTimestamp="2025-10-04 05:58:34 +0000 UTC" firstStartedPulling="2025-10-04 05:58:35.216568674 +0000 UTC m=+10526.508578386" lastFinishedPulling="2025-10-04 05:59:18.111598097 +0000 UTC m=+10569.403607819" observedRunningTime="2025-10-04 05:59:19.166388109 +0000 UTC m=+10570.458397821" watchObservedRunningTime="2025-10-04 05:59:19.183708769 +0000 UTC m=+10570.475718481" Oct 04 05:59:29 crc kubenswrapper[4770]: I1004 05:59:29.685724 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:59:29 crc kubenswrapper[4770]: E1004 05:59:29.686632 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:59:40 crc kubenswrapper[4770]: I1004 05:59:40.673675 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:59:40 crc kubenswrapper[4770]: E1004 05:59:40.674375 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 05:59:51 crc kubenswrapper[4770]: I1004 05:59:51.674601 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 05:59:51 crc kubenswrapper[4770]: E1004 05:59:51.675364 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.188903 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng"] Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.191228 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.193637 4770 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.195777 4770 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.201908 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng"] Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.277570 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-config-volume\") pod \"collect-profiles-29325960-rsqng\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.277654 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twvcx\" (UniqueName: \"kubernetes.io/projected/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-kube-api-access-twvcx\") pod \"collect-profiles-29325960-rsqng\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.277992 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-secret-volume\") pod \"collect-profiles-29325960-rsqng\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.380359 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-config-volume\") pod \"collect-profiles-29325960-rsqng\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.380452 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twvcx\" (UniqueName: \"kubernetes.io/projected/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-kube-api-access-twvcx\") pod \"collect-profiles-29325960-rsqng\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.380584 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-secret-volume\") pod \"collect-profiles-29325960-rsqng\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.382062 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-config-volume\") pod \"collect-profiles-29325960-rsqng\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.399864 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twvcx\" (UniqueName: \"kubernetes.io/projected/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-kube-api-access-twvcx\") pod \"collect-profiles-29325960-rsqng\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.408147 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-secret-volume\") pod \"collect-profiles-29325960-rsqng\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.522619 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:00 crc kubenswrapper[4770]: I1004 06:00:00.830735 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_093c0745-f3d5-4954-9617-f69e6b4a5529/init-config-reloader/0.log" Oct 04 06:00:01 crc kubenswrapper[4770]: I1004 06:00:01.057934 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng"] Oct 04 06:00:01 crc kubenswrapper[4770]: I1004 06:00:01.142407 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_093c0745-f3d5-4954-9617-f69e6b4a5529/alertmanager/0.log" Oct 04 06:00:01 crc kubenswrapper[4770]: I1004 06:00:01.230046 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_093c0745-f3d5-4954-9617-f69e6b4a5529/init-config-reloader/0.log" Oct 04 06:00:01 crc kubenswrapper[4770]: I1004 06:00:01.321330 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_093c0745-f3d5-4954-9617-f69e6b4a5529/config-reloader/0.log" Oct 04 06:00:01 crc kubenswrapper[4770]: I1004 06:00:01.542232 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_6f854c7c-1a25-48b9-9bdc-78b4d7ff550e/aodh-api/0.log" Oct 04 06:00:01 crc kubenswrapper[4770]: I1004 06:00:01.571057 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" event={"ID":"040ce71b-6b6a-45ae-a23a-8d38c37fe4de","Type":"ContainerStarted","Data":"0b0bf927cafac8956ed860f535ddeaaaa042444fd47f65064b3e88fb024f1d20"} Oct 04 06:00:01 crc kubenswrapper[4770]: I1004 06:00:01.571103 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" event={"ID":"040ce71b-6b6a-45ae-a23a-8d38c37fe4de","Type":"ContainerStarted","Data":"f550aee2a557253c7160b5c12f60bf9ee9045f71a15b8fee99852f3cc591b45e"} Oct 04 06:00:01 crc kubenswrapper[4770]: I1004 06:00:01.591482 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" podStartSLOduration=1.591460648 podStartE2EDuration="1.591460648s" podCreationTimestamp="2025-10-04 06:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 06:00:01.587909145 +0000 UTC m=+10612.879918857" watchObservedRunningTime="2025-10-04 06:00:01.591460648 +0000 UTC m=+10612.883470360" Oct 04 06:00:02 crc kubenswrapper[4770]: I1004 06:00:02.428735 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_6f854c7c-1a25-48b9-9bdc-78b4d7ff550e/aodh-evaluator/0.log" Oct 04 06:00:02 crc kubenswrapper[4770]: I1004 06:00:02.437414 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_6f854c7c-1a25-48b9-9bdc-78b4d7ff550e/aodh-notifier/0.log" Oct 04 06:00:02 crc kubenswrapper[4770]: I1004 06:00:02.590972 4770 generic.go:334] "Generic (PLEG): container finished" podID="040ce71b-6b6a-45ae-a23a-8d38c37fe4de" containerID="0b0bf927cafac8956ed860f535ddeaaaa042444fd47f65064b3e88fb024f1d20" exitCode=0 Oct 04 06:00:02 crc kubenswrapper[4770]: I1004 06:00:02.591046 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" event={"ID":"040ce71b-6b6a-45ae-a23a-8d38c37fe4de","Type":"ContainerDied","Data":"0b0bf927cafac8956ed860f535ddeaaaa042444fd47f65064b3e88fb024f1d20"} Oct 04 06:00:02 crc kubenswrapper[4770]: I1004 06:00:02.654738 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_6f854c7c-1a25-48b9-9bdc-78b4d7ff550e/aodh-listener/0.log" Oct 04 06:00:02 crc kubenswrapper[4770]: I1004 06:00:02.657925 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54dd7bd464-g9bnt_dedbb8f2-13d7-4969-9e03-f24f613f0f0a/barbican-api/0.log" Oct 04 06:00:02 crc kubenswrapper[4770]: I1004 06:00:02.817800 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54dd7bd464-g9bnt_dedbb8f2-13d7-4969-9e03-f24f613f0f0a/barbican-api-log/0.log" Oct 04 06:00:02 crc kubenswrapper[4770]: I1004 06:00:02.878350 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-665bfd56dd-lf8wl_d4839fba-4fcd-4b16-a38a-af41ed293efc/barbican-keystone-listener/0.log" Oct 04 06:00:03 crc kubenswrapper[4770]: I1004 06:00:03.015302 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-665bfd56dd-lf8wl_d4839fba-4fcd-4b16-a38a-af41ed293efc/barbican-keystone-listener-log/0.log" Oct 04 06:00:03 crc kubenswrapper[4770]: I1004 06:00:03.087702 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5fc6d5ddf-dhcsd_c73f3d35-51e8-4ad9-be44-78ae4cd1fec1/barbican-worker/0.log" Oct 04 06:00:03 crc kubenswrapper[4770]: I1004 06:00:03.243876 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5fc6d5ddf-dhcsd_c73f3d35-51e8-4ad9-be44-78ae4cd1fec1/barbican-worker-log/0.log" Oct 04 06:00:03 crc kubenswrapper[4770]: I1004 06:00:03.323251 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-7gd5m_55f43d2c-35b3-491a-a080-dee559ff5e84/bootstrap-openstack-openstack-cell1/0.log" Oct 04 06:00:03 crc kubenswrapper[4770]: I1004 06:00:03.524233 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c7631708-8285-44c1-b416-ad9e0dae57b8/ceilometer-notification-agent/0.log" Oct 04 06:00:03 crc kubenswrapper[4770]: I1004 06:00:03.575863 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c7631708-8285-44c1-b416-ad9e0dae57b8/ceilometer-central-agent/0.log" Oct 04 06:00:03 crc kubenswrapper[4770]: I1004 06:00:03.696748 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c7631708-8285-44c1-b416-ad9e0dae57b8/proxy-httpd/0.log" Oct 04 06:00:03 crc kubenswrapper[4770]: I1004 06:00:03.778328 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c7631708-8285-44c1-b416-ad9e0dae57b8/sg-core/0.log" Oct 04 06:00:03 crc kubenswrapper[4770]: I1004 06:00:03.906461 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-4zjfm_654841f2-f5ef-4b6a-bc0f-d78b8dbe5347/ceph-client-openstack-openstack-cell1/0.log" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.032489 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.129284 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c3f8f6e6-6f18-40fc-842b-73323e55e559/cinder-api-log/0.log" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.170332 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c3f8f6e6-6f18-40fc-842b-73323e55e559/cinder-api/0.log" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.199548 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-secret-volume\") pod \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.199661 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twvcx\" (UniqueName: \"kubernetes.io/projected/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-kube-api-access-twvcx\") pod \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.199743 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-config-volume\") pod \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\" (UID: \"040ce71b-6b6a-45ae-a23a-8d38c37fe4de\") " Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.200455 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-config-volume" (OuterVolumeSpecName: "config-volume") pod "040ce71b-6b6a-45ae-a23a-8d38c37fe4de" (UID: "040ce71b-6b6a-45ae-a23a-8d38c37fe4de"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.206368 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "040ce71b-6b6a-45ae-a23a-8d38c37fe4de" (UID: "040ce71b-6b6a-45ae-a23a-8d38c37fe4de"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.209435 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-kube-api-access-twvcx" (OuterVolumeSpecName: "kube-api-access-twvcx") pod "040ce71b-6b6a-45ae-a23a-8d38c37fe4de" (UID: "040ce71b-6b6a-45ae-a23a-8d38c37fe4de"). InnerVolumeSpecName "kube-api-access-twvcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.302701 4770 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.302747 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twvcx\" (UniqueName: \"kubernetes.io/projected/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-kube-api-access-twvcx\") on node \"crc\" DevicePath \"\"" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.302762 4770 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/040ce71b-6b6a-45ae-a23a-8d38c37fe4de-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.504772 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5cbf3298-0dbe-4979-a94f-a6baabc7b545/probe/0.log" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.532076 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5cbf3298-0dbe-4979-a94f-a6baabc7b545/cinder-backup/0.log" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.641685 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" event={"ID":"040ce71b-6b6a-45ae-a23a-8d38c37fe4de","Type":"ContainerDied","Data":"f550aee2a557253c7160b5c12f60bf9ee9045f71a15b8fee99852f3cc591b45e"} Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.641722 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f550aee2a557253c7160b5c12f60bf9ee9045f71a15b8fee99852f3cc591b45e" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.641780 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325960-rsqng" Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.767213 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz"] Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.789226 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-s6wnz"] Oct 04 06:00:04 crc kubenswrapper[4770]: I1004 06:00:04.935088 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3a2e4f90-e9b4-44ed-abef-2de3454b9260/cinder-scheduler/0.log" Oct 04 06:00:05 crc kubenswrapper[4770]: I1004 06:00:05.100124 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3a2e4f90-e9b4-44ed-abef-2de3454b9260/probe/0.log" Oct 04 06:00:05 crc kubenswrapper[4770]: I1004 06:00:05.160717 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_328aff13-c574-4896-b312-cb36ac8315c2/cinder-volume/0.log" Oct 04 06:00:05 crc kubenswrapper[4770]: I1004 06:00:05.475565 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_328aff13-c574-4896-b312-cb36ac8315c2/probe/0.log" Oct 04 06:00:05 crc kubenswrapper[4770]: I1004 06:00:05.692020 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc610f59-e16b-48b2-aef3-fee7f3a76cb1" path="/var/lib/kubelet/pods/fc610f59-e16b-48b2-aef3-fee7f3a76cb1/volumes" Oct 04 06:00:05 crc kubenswrapper[4770]: I1004 06:00:05.895364 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-49sv8_e23f70c8-5f06-4908-9907-f8af47aef701/configure-network-openstack-openstack-cell1/0.log" Oct 04 06:00:05 crc kubenswrapper[4770]: I1004 06:00:05.908788 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-dv8tn_1ceac293-7f9f-45c4-a92a-8c3e82c4c85e/configure-os-openstack-openstack-cell1/0.log" Oct 04 06:00:06 crc kubenswrapper[4770]: I1004 06:00:06.181712 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-f684d669f-jkbvx_cc184b0a-0bf4-4a76-b428-02d3e66e7eb2/init/0.log" Oct 04 06:00:06 crc kubenswrapper[4770]: I1004 06:00:06.364744 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-f684d669f-jkbvx_cc184b0a-0bf4-4a76-b428-02d3e66e7eb2/init/0.log" Oct 04 06:00:06 crc kubenswrapper[4770]: I1004 06:00:06.598934 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-f684d669f-jkbvx_cc184b0a-0bf4-4a76-b428-02d3e66e7eb2/dnsmasq-dns/0.log" Oct 04 06:00:06 crc kubenswrapper[4770]: I1004 06:00:06.675031 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 06:00:06 crc kubenswrapper[4770]: E1004 06:00:06.675661 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:00:06 crc kubenswrapper[4770]: I1004 06:00:06.849529 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-kxtjn_ef4ea9df-722f-4700-b51d-17693400e22a/download-cache-openstack-openstack-cell1/0.log" Oct 04 06:00:07 crc kubenswrapper[4770]: I1004 06:00:07.082552 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6b5e1455-d5f5-49d4-84a5-2aadce69bc3e/glance-log/0.log" Oct 04 06:00:07 crc kubenswrapper[4770]: I1004 06:00:07.132876 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6b5e1455-d5f5-49d4-84a5-2aadce69bc3e/glance-httpd/0.log" Oct 04 06:00:07 crc kubenswrapper[4770]: I1004 06:00:07.342961 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_be23d13e-755b-4ad3-883c-e52f4bb1146c/glance-httpd/0.log" Oct 04 06:00:07 crc kubenswrapper[4770]: I1004 06:00:07.370799 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_be23d13e-755b-4ad3-883c-e52f4bb1146c/glance-log/0.log" Oct 04 06:00:08 crc kubenswrapper[4770]: I1004 06:00:08.042992 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-755fc856cf-4s9v6_58ed5893-3dd0-42bf-b6ea-f2ff90808c98/heat-api/0.log" Oct 04 06:00:08 crc kubenswrapper[4770]: I1004 06:00:08.246207 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-764bcdf645-85424_b5a4f316-6949-4bb8-9192-30adda4c945e/heat-engine/0.log" Oct 04 06:00:08 crc kubenswrapper[4770]: I1004 06:00:08.271057 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5b95c7f4c5-559tc_dd09edfd-2fa6-4516-b3d6-8937cab9ff76/heat-cfnapi/0.log" Oct 04 06:00:09 crc kubenswrapper[4770]: I1004 06:00:09.059810 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-vklws_3cffb45c-2e95-4e28-9c87-56593c1611df/install-certs-openstack-openstack-cell1/0.log" Oct 04 06:00:09 crc kubenswrapper[4770]: I1004 06:00:09.155297 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64f6748dff-9x2w7_025aaf1a-ee7f-4715-8f03-dadd939d8bda/horizon/0.log" Oct 04 06:00:09 crc kubenswrapper[4770]: I1004 06:00:09.165716 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64f6748dff-9x2w7_025aaf1a-ee7f-4715-8f03-dadd939d8bda/horizon-log/0.log" Oct 04 06:00:09 crc kubenswrapper[4770]: I1004 06:00:09.396847 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-zdb69_8b6ca520-b8fe-491e-8a3d-307a5489c908/install-os-openstack-openstack-cell1/0.log" Oct 04 06:00:09 crc kubenswrapper[4770]: I1004 06:00:09.776099 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5454b8cff9-nj62p_15d1751b-3112-4f88-ab65-8bcf99e6e336/keystone-api/0.log" Oct 04 06:00:10 crc kubenswrapper[4770]: I1004 06:00:10.250521 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a9e94c14-ec90-4ff2-97a8-8b98ea370af1/kube-state-metrics/0.log" Oct 04 06:00:10 crc kubenswrapper[4770]: I1004 06:00:10.265027 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325901-8rxql_ff136520-ed4e-478b-9ac5-69b54366cf0e/keystone-cron/0.log" Oct 04 06:00:10 crc kubenswrapper[4770]: I1004 06:00:10.473263 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-zhbgq_dab7ef71-5add-4bff-b920-a2b2a974441a/libvirt-openstack-openstack-cell1/0.log" Oct 04 06:00:10 crc kubenswrapper[4770]: I1004 06:00:10.737148 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc/manila-api/0.log" Oct 04 06:00:10 crc kubenswrapper[4770]: I1004 06:00:10.794193 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_5a9b426b-51b2-4e4f-bcb8-c0b84d6313dc/manila-api-log/0.log" Oct 04 06:00:11 crc kubenswrapper[4770]: I1004 06:00:11.375963 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_7e858e72-afbf-45b7-96fb-c22def95cbed/probe/0.log" Oct 04 06:00:11 crc kubenswrapper[4770]: I1004 06:00:11.776375 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_7e858e72-afbf-45b7-96fb-c22def95cbed/manila-scheduler/0.log" Oct 04 06:00:12 crc kubenswrapper[4770]: I1004 06:00:12.296665 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_e74b101f-cfa1-4c29-b78a-3c48964a271b/probe/0.log" Oct 04 06:00:12 crc kubenswrapper[4770]: I1004 06:00:12.649167 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_e74b101f-cfa1-4c29-b78a-3c48964a271b/manila-share/0.log" Oct 04 06:00:13 crc kubenswrapper[4770]: I1004 06:00:13.185602 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7ddcc967fc-sj8jm_ea02a541-62ae-41c6-bfde-6b64f6081e8c/neutron-httpd/0.log" Oct 04 06:00:13 crc kubenswrapper[4770]: I1004 06:00:13.363585 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7ddcc967fc-sj8jm_ea02a541-62ae-41c6-bfde-6b64f6081e8c/neutron-api/0.log" Oct 04 06:00:15 crc kubenswrapper[4770]: I1004 06:00:15.757737 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-68ssf_6707d041-c3e2-4445-b381-ce39c1e572a8/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 04 06:00:15 crc kubenswrapper[4770]: I1004 06:00:15.946682 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-stt44_c854c1c1-d5eb-471c-9c90-755b2514480c/neutron-metadata-openstack-openstack-cell1/0.log" Oct 04 06:00:16 crc kubenswrapper[4770]: I1004 06:00:16.969053 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fbff40a9-a564-4416-8e59-3cc50cef4b90/nova-api-api/0.log" Oct 04 06:00:17 crc kubenswrapper[4770]: I1004 06:00:17.194707 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-gjmnf_ee2615a0-868c-4c77-ae4c-cfc101b62266/neutron-sriov-openstack-openstack-cell1/0.log" Oct 04 06:00:17 crc kubenswrapper[4770]: I1004 06:00:17.296122 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fbff40a9-a564-4416-8e59-3cc50cef4b90/nova-api-log/0.log" Oct 04 06:00:17 crc kubenswrapper[4770]: I1004 06:00:17.342700 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_b430eca9-a710-4be3-b5ee-228951137298/memcached/0.log" Oct 04 06:00:17 crc kubenswrapper[4770]: I1004 06:00:17.590702 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_28662e31-5054-4536-84c7-e22941f7dc19/nova-cell0-conductor-conductor/0.log" Oct 04 06:00:17 crc kubenswrapper[4770]: I1004 06:00:17.895074 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_6e87d157-36eb-4197-a066-cefba0bddb76/nova-cell1-conductor-conductor/0.log" Oct 04 06:00:17 crc kubenswrapper[4770]: I1004 06:00:17.929892 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_481d5341-bd64-43ad-be7d-54ce70e96f29/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 06:00:18 crc kubenswrapper[4770]: I1004 06:00:18.976629 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c3ce320f-ae54-4e20-915d-c1c1a9a1c887/nova-metadata-log/0.log" Oct 04 06:00:19 crc kubenswrapper[4770]: I1004 06:00:19.174818 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellrfvnw_9549539a-734a-482a-a374-fbd7be4dc5cc/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 04 06:00:19 crc kubenswrapper[4770]: I1004 06:00:19.468658 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c3ce320f-ae54-4e20-915d-c1c1a9a1c887/nova-metadata-metadata/0.log" Oct 04 06:00:20 crc kubenswrapper[4770]: I1004 06:00:20.060747 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5968f659f7-n2kjd_198ec713-5ae0-4e0a-ba62-433810e9e486/init/0.log" Oct 04 06:00:20 crc kubenswrapper[4770]: I1004 06:00:20.116558 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-5kbhh_3497238c-e3ec-4169-a636-342eca9f2b8e/nova-cell1-openstack-openstack-cell1/0.log" Oct 04 06:00:20 crc kubenswrapper[4770]: I1004 06:00:20.149414 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ca0f0db0-e939-4054-9aa4-8d610293082f/nova-scheduler-scheduler/0.log" Oct 04 06:00:20 crc kubenswrapper[4770]: I1004 06:00:20.674104 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 06:00:20 crc kubenswrapper[4770]: E1004 06:00:20.674441 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:00:20 crc kubenswrapper[4770]: I1004 06:00:20.862290 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5968f659f7-n2kjd_198ec713-5ae0-4e0a-ba62-433810e9e486/init/0.log" Oct 04 06:00:21 crc kubenswrapper[4770]: I1004 06:00:21.177628 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5968f659f7-n2kjd_198ec713-5ae0-4e0a-ba62-433810e9e486/octavia-api-provider-agent/0.log" Oct 04 06:00:21 crc kubenswrapper[4770]: I1004 06:00:21.417302 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5968f659f7-n2kjd_198ec713-5ae0-4e0a-ba62-433810e9e486/octavia-api/0.log" Oct 04 06:00:21 crc kubenswrapper[4770]: I1004 06:00:21.546852 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-47zrr_67bf6fd1-b5eb-4c67-8210-0839f7cdb14b/init/0.log" Oct 04 06:00:21 crc kubenswrapper[4770]: I1004 06:00:21.806526 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-47zrr_67bf6fd1-b5eb-4c67-8210-0839f7cdb14b/octavia-healthmanager/0.log" Oct 04 06:00:21 crc kubenswrapper[4770]: I1004 06:00:21.828358 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-47zrr_67bf6fd1-b5eb-4c67-8210-0839f7cdb14b/init/0.log" Oct 04 06:00:22 crc kubenswrapper[4770]: I1004 06:00:22.079213 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-wx6tq_aadc601e-a24b-4258-91c5-c1f703ed1973/init/0.log" Oct 04 06:00:23 crc kubenswrapper[4770]: I1004 06:00:23.060274 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-wx6tq_aadc601e-a24b-4258-91c5-c1f703ed1973/octavia-housekeeping/0.log" Oct 04 06:00:23 crc kubenswrapper[4770]: I1004 06:00:23.180807 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-wx6tq_aadc601e-a24b-4258-91c5-c1f703ed1973/init/0.log" Oct 04 06:00:23 crc kubenswrapper[4770]: I1004 06:00:23.400695 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-678599687f-jn5sj_f4c708b4-49e9-4a43-9372-c932bb4bc473/init/0.log" Oct 04 06:00:23 crc kubenswrapper[4770]: I1004 06:00:23.400983 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-678599687f-jn5sj_f4c708b4-49e9-4a43-9372-c932bb4bc473/init/0.log" Oct 04 06:00:23 crc kubenswrapper[4770]: I1004 06:00:23.442852 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-678599687f-jn5sj_f4c708b4-49e9-4a43-9372-c932bb4bc473/octavia-amphora-httpd/0.log" Oct 04 06:00:23 crc kubenswrapper[4770]: I1004 06:00:23.633091 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-jxh7k_c388dfff-1c4e-4b28-9998-edea3809767b/init/0.log" Oct 04 06:00:23 crc kubenswrapper[4770]: I1004 06:00:23.768833 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-jxh7k_c388dfff-1c4e-4b28-9998-edea3809767b/init/0.log" Oct 04 06:00:23 crc kubenswrapper[4770]: I1004 06:00:23.817796 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-jxh7k_c388dfff-1c4e-4b28-9998-edea3809767b/octavia-rsyslog/0.log" Oct 04 06:00:24 crc kubenswrapper[4770]: I1004 06:00:24.039438 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-d69sh_11a80330-ab5a-4f00-b3b4-341d6d23a268/init/0.log" Oct 04 06:00:24 crc kubenswrapper[4770]: I1004 06:00:24.106851 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-d69sh_11a80330-ab5a-4f00-b3b4-341d6d23a268/init/0.log" Oct 04 06:00:24 crc kubenswrapper[4770]: I1004 06:00:24.359976 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e130aa71-bdc5-49f4-8fa4-d2e96cbf359e/mysql-bootstrap/0.log" Oct 04 06:00:24 crc kubenswrapper[4770]: I1004 06:00:24.517310 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-d69sh_11a80330-ab5a-4f00-b3b4-341d6d23a268/octavia-worker/0.log" Oct 04 06:00:24 crc kubenswrapper[4770]: I1004 06:00:24.680337 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e130aa71-bdc5-49f4-8fa4-d2e96cbf359e/mysql-bootstrap/0.log" Oct 04 06:00:24 crc kubenswrapper[4770]: I1004 06:00:24.759943 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e130aa71-bdc5-49f4-8fa4-d2e96cbf359e/galera/0.log" Oct 04 06:00:24 crc kubenswrapper[4770]: I1004 06:00:24.949475 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3bceaca0-f90b-4feb-b54c-6292db9d580c/mysql-bootstrap/0.log" Oct 04 06:00:25 crc kubenswrapper[4770]: I1004 06:00:25.107205 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3bceaca0-f90b-4feb-b54c-6292db9d580c/mysql-bootstrap/0.log" Oct 04 06:00:25 crc kubenswrapper[4770]: I1004 06:00:25.147564 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3bceaca0-f90b-4feb-b54c-6292db9d580c/galera/0.log" Oct 04 06:00:25 crc kubenswrapper[4770]: I1004 06:00:25.328265 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_97cf3f4a-91b4-4e54-ab01-ec01f4ea2d41/openstackclient/0.log" Oct 04 06:00:25 crc kubenswrapper[4770]: I1004 06:00:25.438867 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-n4zz4_5fcae710-09dc-499b-8bba-fcab1d2b6811/openstack-network-exporter/0.log" Oct 04 06:00:25 crc kubenswrapper[4770]: I1004 06:00:25.638008 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ncsvk_8e6ae49d-7530-44a5-bfb1-6201c9e745c8/ovn-controller/0.log" Oct 04 06:00:25 crc kubenswrapper[4770]: I1004 06:00:25.985405 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rwzj2_ea243e32-f3d6-47a6-b2fe-9728af6bc06a/ovsdb-server-init/0.log" Oct 04 06:00:26 crc kubenswrapper[4770]: I1004 06:00:26.445264 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rwzj2_ea243e32-f3d6-47a6-b2fe-9728af6bc06a/ovsdb-server-init/0.log" Oct 04 06:00:26 crc kubenswrapper[4770]: I1004 06:00:26.501434 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rwzj2_ea243e32-f3d6-47a6-b2fe-9728af6bc06a/ovsdb-server/0.log" Oct 04 06:00:26 crc kubenswrapper[4770]: I1004 06:00:26.517866 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rwzj2_ea243e32-f3d6-47a6-b2fe-9728af6bc06a/ovs-vswitchd/0.log" Oct 04 06:00:26 crc kubenswrapper[4770]: I1004 06:00:26.697412 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fbbe2b0e-dad0-4407-9347-59447104e584/openstack-network-exporter/0.log" Oct 04 06:00:26 crc kubenswrapper[4770]: I1004 06:00:26.771204 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fbbe2b0e-dad0-4407-9347-59447104e584/ovn-northd/0.log" Oct 04 06:00:27 crc kubenswrapper[4770]: I1004 06:00:27.203734 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6254f2df-3ad1-4170-abbe-ba955e31bb82/openstack-network-exporter/0.log" Oct 04 06:00:27 crc kubenswrapper[4770]: I1004 06:00:27.465126 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6254f2df-3ad1-4170-abbe-ba955e31bb82/ovsdbserver-nb/0.log" Oct 04 06:00:27 crc kubenswrapper[4770]: I1004 06:00:27.681700 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_3aaa3b43-6349-4b31-a8fd-25bffad0fb69/openstack-network-exporter/0.log" Oct 04 06:00:27 crc kubenswrapper[4770]: I1004 06:00:27.781969 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_3aaa3b43-6349-4b31-a8fd-25bffad0fb69/ovsdbserver-nb/0.log" Oct 04 06:00:28 crc kubenswrapper[4770]: I1004 06:00:28.020147 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_df7e5b43-d3ed-42c0-8656-23460260ad8b/openstack-network-exporter/0.log" Oct 04 06:00:28 crc kubenswrapper[4770]: I1004 06:00:28.152703 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_df7e5b43-d3ed-42c0-8656-23460260ad8b/ovsdbserver-nb/0.log" Oct 04 06:00:28 crc kubenswrapper[4770]: I1004 06:00:28.159131 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-95slf_535f0f1c-4b20-45bf-b2e2-d16481bfd7c1/ovn-openstack-openstack-cell1/0.log" Oct 04 06:00:28 crc kubenswrapper[4770]: I1004 06:00:28.343441 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_417c899d-895d-45ee-b4f2-5730df096041/openstack-network-exporter/0.log" Oct 04 06:00:28 crc kubenswrapper[4770]: I1004 06:00:28.457344 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_417c899d-895d-45ee-b4f2-5730df096041/ovsdbserver-sb/0.log" Oct 04 06:00:28 crc kubenswrapper[4770]: I1004 06:00:28.540597 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_8c885a90-64e9-4c59-9e08-fbbfb07656ab/openstack-network-exporter/0.log" Oct 04 06:00:28 crc kubenswrapper[4770]: I1004 06:00:28.907534 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_e7e0e8df-5568-462a-b40b-0ffc03b6197e/openstack-network-exporter/0.log" Oct 04 06:00:29 crc kubenswrapper[4770]: I1004 06:00:29.185095 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_8c885a90-64e9-4c59-9e08-fbbfb07656ab/ovsdbserver-sb/0.log" Oct 04 06:00:29 crc kubenswrapper[4770]: I1004 06:00:29.339330 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_e7e0e8df-5568-462a-b40b-0ffc03b6197e/ovsdbserver-sb/0.log" Oct 04 06:00:29 crc kubenswrapper[4770]: I1004 06:00:29.465789 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-b475498cb-2njbg_ced10f27-425f-4a56-9d11-9bce2a12a8ac/placement-api/0.log" Oct 04 06:00:29 crc kubenswrapper[4770]: I1004 06:00:29.534639 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-b475498cb-2njbg_ced10f27-425f-4a56-9d11-9bce2a12a8ac/placement-log/0.log" Oct 04 06:00:29 crc kubenswrapper[4770]: I1004 06:00:29.795163 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-chdgtx_94268810-ff27-479b-981f-4dd1f3b52099/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 04 06:00:30 crc kubenswrapper[4770]: I1004 06:00:30.099460 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_00a56b1a-324e-4d06-a234-0821994a1a48/init-config-reloader/0.log" Oct 04 06:00:30 crc kubenswrapper[4770]: I1004 06:00:30.818337 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_00a56b1a-324e-4d06-a234-0821994a1a48/config-reloader/0.log" Oct 04 06:00:30 crc kubenswrapper[4770]: I1004 06:00:30.880688 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_00a56b1a-324e-4d06-a234-0821994a1a48/prometheus/0.log" Oct 04 06:00:31 crc kubenswrapper[4770]: I1004 06:00:31.172437 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_00a56b1a-324e-4d06-a234-0821994a1a48/init-config-reloader/0.log" Oct 04 06:00:31 crc kubenswrapper[4770]: I1004 06:00:31.197057 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_00a56b1a-324e-4d06-a234-0821994a1a48/thanos-sidecar/0.log" Oct 04 06:00:31 crc kubenswrapper[4770]: I1004 06:00:31.413171 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_60cb9246-1987-4782-97e3-84935639077f/setup-container/0.log" Oct 04 06:00:31 crc kubenswrapper[4770]: I1004 06:00:31.649885 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_60cb9246-1987-4782-97e3-84935639077f/setup-container/0.log" Oct 04 06:00:31 crc kubenswrapper[4770]: I1004 06:00:31.802352 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_60cb9246-1987-4782-97e3-84935639077f/rabbitmq/0.log" Oct 04 06:00:32 crc kubenswrapper[4770]: I1004 06:00:32.488176 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b8671f84-8a47-4d20-8272-90b7fd4d6b3d/setup-container/0.log" Oct 04 06:00:32 crc kubenswrapper[4770]: I1004 06:00:32.676827 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b8671f84-8a47-4d20-8272-90b7fd4d6b3d/setup-container/0.log" Oct 04 06:00:33 crc kubenswrapper[4770]: I1004 06:00:33.059374 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-2v52b_704ceb18-dc32-4a0d-ad0a-dee111bae162/reboot-os-openstack-openstack-cell1/0.log" Oct 04 06:00:33 crc kubenswrapper[4770]: I1004 06:00:33.275669 4770 scope.go:117] "RemoveContainer" containerID="90cc942168d13b57d964b1830177602a5b512b28713b93edf5b6ac523bf0aa40" Oct 04 06:00:33 crc kubenswrapper[4770]: I1004 06:00:33.321580 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-p7q9s_a8ebb9de-f667-40ba-9d22-897cbabb320e/run-os-openstack-openstack-cell1/0.log" Oct 04 06:00:33 crc kubenswrapper[4770]: I1004 06:00:33.674479 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 06:00:33 crc kubenswrapper[4770]: E1004 06:00:33.675254 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:00:33 crc kubenswrapper[4770]: I1004 06:00:33.690256 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-rz6bz_13e4633f-4f16-4e0c-8cfb-b500a9ba63f1/ssh-known-hosts-openstack/0.log" Oct 04 06:00:34 crc kubenswrapper[4770]: I1004 06:00:34.098572 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-bw8b9_a84ee920-d896-418c-afe3-91d4e590e350/telemetry-openstack-openstack-cell1/0.log" Oct 04 06:00:34 crc kubenswrapper[4770]: I1004 06:00:34.661271 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-2r848_b5dcbbe1-7d4d-4c82-a705-f367cb129883/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 04 06:00:34 crc kubenswrapper[4770]: I1004 06:00:34.867842 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-64fpm_17d78a1f-a6ab-4523-a033-f9388abb2ada/validate-network-openstack-openstack-cell1/0.log" Oct 04 06:00:35 crc kubenswrapper[4770]: I1004 06:00:35.673526 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b8671f84-8a47-4d20-8272-90b7fd4d6b3d/rabbitmq/0.log" Oct 04 06:00:48 crc kubenswrapper[4770]: I1004 06:00:48.674130 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 06:00:48 crc kubenswrapper[4770]: E1004 06:00:48.674929 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.155059 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325961-rc7v8"] Oct 04 06:01:00 crc kubenswrapper[4770]: E1004 06:01:00.158177 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="040ce71b-6b6a-45ae-a23a-8d38c37fe4de" containerName="collect-profiles" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.158197 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="040ce71b-6b6a-45ae-a23a-8d38c37fe4de" containerName="collect-profiles" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.158473 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="040ce71b-6b6a-45ae-a23a-8d38c37fe4de" containerName="collect-profiles" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.159340 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.195753 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325961-rc7v8"] Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.292418 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-config-data\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.292539 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-fernet-keys\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.292599 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2xtp\" (UniqueName: \"kubernetes.io/projected/9cf4ca9b-edd8-49f0-ab44-e226be692943-kube-api-access-x2xtp\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.292757 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-combined-ca-bundle\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.394881 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-combined-ca-bundle\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.395025 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-config-data\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.395100 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-fernet-keys\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.395157 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2xtp\" (UniqueName: \"kubernetes.io/projected/9cf4ca9b-edd8-49f0-ab44-e226be692943-kube-api-access-x2xtp\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.445738 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-fernet-keys\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.445758 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-combined-ca-bundle\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.450998 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2xtp\" (UniqueName: \"kubernetes.io/projected/9cf4ca9b-edd8-49f0-ab44-e226be692943-kube-api-access-x2xtp\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.460590 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-config-data\") pod \"keystone-cron-29325961-rc7v8\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:00 crc kubenswrapper[4770]: I1004 06:01:00.492831 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:01 crc kubenswrapper[4770]: I1004 06:01:01.033348 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325961-rc7v8"] Oct 04 06:01:01 crc kubenswrapper[4770]: I1004 06:01:01.338263 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325961-rc7v8" event={"ID":"9cf4ca9b-edd8-49f0-ab44-e226be692943","Type":"ContainerStarted","Data":"1f9b5b178db5a244ece9c7cec685f1e422c3a888446b86f15c8eccb06e79bb00"} Oct 04 06:01:01 crc kubenswrapper[4770]: I1004 06:01:01.673855 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 06:01:01 crc kubenswrapper[4770]: E1004 06:01:01.674355 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:01:02 crc kubenswrapper[4770]: I1004 06:01:02.351408 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325961-rc7v8" event={"ID":"9cf4ca9b-edd8-49f0-ab44-e226be692943","Type":"ContainerStarted","Data":"43cdf2f5330b9d841dd8b4579eab28cebca01696e75a7e61600c4a7bc3a8a48b"} Oct 04 06:01:02 crc kubenswrapper[4770]: I1004 06:01:02.378599 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325961-rc7v8" podStartSLOduration=2.378580136 podStartE2EDuration="2.378580136s" podCreationTimestamp="2025-10-04 06:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 06:01:02.367097721 +0000 UTC m=+10673.659107433" watchObservedRunningTime="2025-10-04 06:01:02.378580136 +0000 UTC m=+10673.670589848" Oct 04 06:01:08 crc kubenswrapper[4770]: I1004 06:01:08.431608 4770 generic.go:334] "Generic (PLEG): container finished" podID="9cf4ca9b-edd8-49f0-ab44-e226be692943" containerID="43cdf2f5330b9d841dd8b4579eab28cebca01696e75a7e61600c4a7bc3a8a48b" exitCode=0 Oct 04 06:01:08 crc kubenswrapper[4770]: I1004 06:01:08.431673 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325961-rc7v8" event={"ID":"9cf4ca9b-edd8-49f0-ab44-e226be692943","Type":"ContainerDied","Data":"43cdf2f5330b9d841dd8b4579eab28cebca01696e75a7e61600c4a7bc3a8a48b"} Oct 04 06:01:09 crc kubenswrapper[4770]: I1004 06:01:09.859406 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.006806 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-config-data\") pod \"9cf4ca9b-edd8-49f0-ab44-e226be692943\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.007581 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-fernet-keys\") pod \"9cf4ca9b-edd8-49f0-ab44-e226be692943\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.007624 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-combined-ca-bundle\") pod \"9cf4ca9b-edd8-49f0-ab44-e226be692943\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.007815 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2xtp\" (UniqueName: \"kubernetes.io/projected/9cf4ca9b-edd8-49f0-ab44-e226be692943-kube-api-access-x2xtp\") pod \"9cf4ca9b-edd8-49f0-ab44-e226be692943\" (UID: \"9cf4ca9b-edd8-49f0-ab44-e226be692943\") " Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.020957 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9cf4ca9b-edd8-49f0-ab44-e226be692943" (UID: "9cf4ca9b-edd8-49f0-ab44-e226be692943"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.032286 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cf4ca9b-edd8-49f0-ab44-e226be692943-kube-api-access-x2xtp" (OuterVolumeSpecName: "kube-api-access-x2xtp") pod "9cf4ca9b-edd8-49f0-ab44-e226be692943" (UID: "9cf4ca9b-edd8-49f0-ab44-e226be692943"). InnerVolumeSpecName "kube-api-access-x2xtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.046115 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cf4ca9b-edd8-49f0-ab44-e226be692943" (UID: "9cf4ca9b-edd8-49f0-ab44-e226be692943"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.069897 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-config-data" (OuterVolumeSpecName: "config-data") pod "9cf4ca9b-edd8-49f0-ab44-e226be692943" (UID: "9cf4ca9b-edd8-49f0-ab44-e226be692943"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.109886 4770 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.109936 4770 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.109949 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2xtp\" (UniqueName: \"kubernetes.io/projected/9cf4ca9b-edd8-49f0-ab44-e226be692943-kube-api-access-x2xtp\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.109962 4770 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cf4ca9b-edd8-49f0-ab44-e226be692943-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.458234 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325961-rc7v8" event={"ID":"9cf4ca9b-edd8-49f0-ab44-e226be692943","Type":"ContainerDied","Data":"1f9b5b178db5a244ece9c7cec685f1e422c3a888446b86f15c8eccb06e79bb00"} Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.458283 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f9b5b178db5a244ece9c7cec685f1e422c3a888446b86f15c8eccb06e79bb00" Oct 04 06:01:10 crc kubenswrapper[4770]: I1004 06:01:10.458301 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325961-rc7v8" Oct 04 06:01:14 crc kubenswrapper[4770]: I1004 06:01:14.674530 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 06:01:14 crc kubenswrapper[4770]: E1004 06:01:14.675423 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:01:27 crc kubenswrapper[4770]: I1004 06:01:27.675273 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 06:01:27 crc kubenswrapper[4770]: E1004 06:01:27.676597 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:01:42 crc kubenswrapper[4770]: I1004 06:01:42.674696 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 06:01:43 crc kubenswrapper[4770]: I1004 06:01:43.820237 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"2281e97bdf52b9438829f5ef88f5250486a7610080a67d0b0aa67a36806cc454"} Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.022146 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s8lt8"] Oct 04 06:02:36 crc kubenswrapper[4770]: E1004 06:02:36.023281 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf4ca9b-edd8-49f0-ab44-e226be692943" containerName="keystone-cron" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.023299 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf4ca9b-edd8-49f0-ab44-e226be692943" containerName="keystone-cron" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.023601 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cf4ca9b-edd8-49f0-ab44-e226be692943" containerName="keystone-cron" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.025357 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.039275 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8lt8"] Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.210194 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-utilities\") pod \"redhat-marketplace-s8lt8\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.210657 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-catalog-content\") pod \"redhat-marketplace-s8lt8\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.210712 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2nxw\" (UniqueName: \"kubernetes.io/projected/8c7910d4-ab13-453e-999e-44dad61a6195-kube-api-access-j2nxw\") pod \"redhat-marketplace-s8lt8\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.313072 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-catalog-content\") pod \"redhat-marketplace-s8lt8\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.313156 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2nxw\" (UniqueName: \"kubernetes.io/projected/8c7910d4-ab13-453e-999e-44dad61a6195-kube-api-access-j2nxw\") pod \"redhat-marketplace-s8lt8\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.313191 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-utilities\") pod \"redhat-marketplace-s8lt8\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.313618 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-catalog-content\") pod \"redhat-marketplace-s8lt8\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.313644 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-utilities\") pod \"redhat-marketplace-s8lt8\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.333598 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2nxw\" (UniqueName: \"kubernetes.io/projected/8c7910d4-ab13-453e-999e-44dad61a6195-kube-api-access-j2nxw\") pod \"redhat-marketplace-s8lt8\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:36 crc kubenswrapper[4770]: I1004 06:02:36.417750 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:37 crc kubenswrapper[4770]: I1004 06:02:37.026037 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8lt8"] Oct 04 06:02:37 crc kubenswrapper[4770]: I1004 06:02:37.397650 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8lt8" event={"ID":"8c7910d4-ab13-453e-999e-44dad61a6195","Type":"ContainerStarted","Data":"a86024dfac9ba8b3a8bc22d55d7c36eec4bee71b925d5dcae4a5eb4ff7dfa229"} Oct 04 06:02:38 crc kubenswrapper[4770]: I1004 06:02:38.410668 4770 generic.go:334] "Generic (PLEG): container finished" podID="8c7910d4-ab13-453e-999e-44dad61a6195" containerID="17fb6fa16512bf2db9426f3319a6f1b3ed7cbdee1c15d331f489208e47aa7f08" exitCode=0 Oct 04 06:02:38 crc kubenswrapper[4770]: I1004 06:02:38.410949 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8lt8" event={"ID":"8c7910d4-ab13-453e-999e-44dad61a6195","Type":"ContainerDied","Data":"17fb6fa16512bf2db9426f3319a6f1b3ed7cbdee1c15d331f489208e47aa7f08"} Oct 04 06:02:40 crc kubenswrapper[4770]: I1004 06:02:40.444939 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8lt8" event={"ID":"8c7910d4-ab13-453e-999e-44dad61a6195","Type":"ContainerStarted","Data":"e597941bca544b04b1b895a7c69ed7428ccc47c173779641cc4ca74219166568"} Oct 04 06:02:42 crc kubenswrapper[4770]: I1004 06:02:42.465637 4770 generic.go:334] "Generic (PLEG): container finished" podID="8c7910d4-ab13-453e-999e-44dad61a6195" containerID="e597941bca544b04b1b895a7c69ed7428ccc47c173779641cc4ca74219166568" exitCode=0 Oct 04 06:02:42 crc kubenswrapper[4770]: I1004 06:02:42.465737 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8lt8" event={"ID":"8c7910d4-ab13-453e-999e-44dad61a6195","Type":"ContainerDied","Data":"e597941bca544b04b1b895a7c69ed7428ccc47c173779641cc4ca74219166568"} Oct 04 06:02:44 crc kubenswrapper[4770]: I1004 06:02:44.489849 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8lt8" event={"ID":"8c7910d4-ab13-453e-999e-44dad61a6195","Type":"ContainerStarted","Data":"3f7040e2a6da53829b35eb259381991ebafbf365b1d9f9667e7e750708c7b25c"} Oct 04 06:02:45 crc kubenswrapper[4770]: I1004 06:02:45.501067 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/crc-debug-zh67x" event={"ID":"06582714-b3e5-4206-9a81-ecb43ec9513d","Type":"ContainerDied","Data":"f2384b679082368582479583a5c36a28ab4d7a321803c54c10c1fdb1166afd84"} Oct 04 06:02:45 crc kubenswrapper[4770]: I1004 06:02:45.501002 4770 generic.go:334] "Generic (PLEG): container finished" podID="06582714-b3e5-4206-9a81-ecb43ec9513d" containerID="f2384b679082368582479583a5c36a28ab4d7a321803c54c10c1fdb1166afd84" exitCode=0 Oct 04 06:02:45 crc kubenswrapper[4770]: I1004 06:02:45.520310 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s8lt8" podStartSLOduration=4.882558235 podStartE2EDuration="10.520294859s" podCreationTimestamp="2025-10-04 06:02:35 +0000 UTC" firstStartedPulling="2025-10-04 06:02:38.413347768 +0000 UTC m=+10769.705357490" lastFinishedPulling="2025-10-04 06:02:44.051084402 +0000 UTC m=+10775.343094114" observedRunningTime="2025-10-04 06:02:45.518534973 +0000 UTC m=+10776.810544695" watchObservedRunningTime="2025-10-04 06:02:45.520294859 +0000 UTC m=+10776.812304561" Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.418864 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.419280 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.469100 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.633374 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.680564 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c885t/crc-debug-zh67x"] Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.693242 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c885t/crc-debug-zh67x"] Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.756769 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lgrr\" (UniqueName: \"kubernetes.io/projected/06582714-b3e5-4206-9a81-ecb43ec9513d-kube-api-access-4lgrr\") pod \"06582714-b3e5-4206-9a81-ecb43ec9513d\" (UID: \"06582714-b3e5-4206-9a81-ecb43ec9513d\") " Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.757091 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06582714-b3e5-4206-9a81-ecb43ec9513d-host\") pod \"06582714-b3e5-4206-9a81-ecb43ec9513d\" (UID: \"06582714-b3e5-4206-9a81-ecb43ec9513d\") " Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.757175 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06582714-b3e5-4206-9a81-ecb43ec9513d-host" (OuterVolumeSpecName: "host") pod "06582714-b3e5-4206-9a81-ecb43ec9513d" (UID: "06582714-b3e5-4206-9a81-ecb43ec9513d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.758484 4770 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06582714-b3e5-4206-9a81-ecb43ec9513d-host\") on node \"crc\" DevicePath \"\"" Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.763651 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06582714-b3e5-4206-9a81-ecb43ec9513d-kube-api-access-4lgrr" (OuterVolumeSpecName: "kube-api-access-4lgrr") pod "06582714-b3e5-4206-9a81-ecb43ec9513d" (UID: "06582714-b3e5-4206-9a81-ecb43ec9513d"). InnerVolumeSpecName "kube-api-access-4lgrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:02:46 crc kubenswrapper[4770]: I1004 06:02:46.860620 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lgrr\" (UniqueName: \"kubernetes.io/projected/06582714-b3e5-4206-9a81-ecb43ec9513d-kube-api-access-4lgrr\") on node \"crc\" DevicePath \"\"" Oct 04 06:02:47 crc kubenswrapper[4770]: I1004 06:02:47.522714 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-zh67x" Oct 04 06:02:47 crc kubenswrapper[4770]: I1004 06:02:47.522712 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f70f3f6154bd59c9cc975ec529c9c849aee681ca03d8c1b66b7c77df65ccb43" Oct 04 06:02:47 crc kubenswrapper[4770]: I1004 06:02:47.695495 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06582714-b3e5-4206-9a81-ecb43ec9513d" path="/var/lib/kubelet/pods/06582714-b3e5-4206-9a81-ecb43ec9513d/volumes" Oct 04 06:02:47 crc kubenswrapper[4770]: I1004 06:02:47.909394 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c885t/crc-debug-vgq7f"] Oct 04 06:02:47 crc kubenswrapper[4770]: E1004 06:02:47.909902 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06582714-b3e5-4206-9a81-ecb43ec9513d" containerName="container-00" Oct 04 06:02:47 crc kubenswrapper[4770]: I1004 06:02:47.909924 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="06582714-b3e5-4206-9a81-ecb43ec9513d" containerName="container-00" Oct 04 06:02:47 crc kubenswrapper[4770]: I1004 06:02:47.910243 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="06582714-b3e5-4206-9a81-ecb43ec9513d" containerName="container-00" Oct 04 06:02:47 crc kubenswrapper[4770]: I1004 06:02:47.911069 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:48 crc kubenswrapper[4770]: I1004 06:02:48.094814 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz4qz\" (UniqueName: \"kubernetes.io/projected/d3fb511c-436f-4480-a590-9f7a609ea562-kube-api-access-vz4qz\") pod \"crc-debug-vgq7f\" (UID: \"d3fb511c-436f-4480-a590-9f7a609ea562\") " pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:48 crc kubenswrapper[4770]: I1004 06:02:48.094921 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3fb511c-436f-4480-a590-9f7a609ea562-host\") pod \"crc-debug-vgq7f\" (UID: \"d3fb511c-436f-4480-a590-9f7a609ea562\") " pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:48 crc kubenswrapper[4770]: I1004 06:02:48.197562 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz4qz\" (UniqueName: \"kubernetes.io/projected/d3fb511c-436f-4480-a590-9f7a609ea562-kube-api-access-vz4qz\") pod \"crc-debug-vgq7f\" (UID: \"d3fb511c-436f-4480-a590-9f7a609ea562\") " pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:48 crc kubenswrapper[4770]: I1004 06:02:48.198497 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3fb511c-436f-4480-a590-9f7a609ea562-host\") pod \"crc-debug-vgq7f\" (UID: \"d3fb511c-436f-4480-a590-9f7a609ea562\") " pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:48 crc kubenswrapper[4770]: I1004 06:02:48.198636 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3fb511c-436f-4480-a590-9f7a609ea562-host\") pod \"crc-debug-vgq7f\" (UID: \"d3fb511c-436f-4480-a590-9f7a609ea562\") " pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:48 crc kubenswrapper[4770]: I1004 06:02:48.221818 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz4qz\" (UniqueName: \"kubernetes.io/projected/d3fb511c-436f-4480-a590-9f7a609ea562-kube-api-access-vz4qz\") pod \"crc-debug-vgq7f\" (UID: \"d3fb511c-436f-4480-a590-9f7a609ea562\") " pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:48 crc kubenswrapper[4770]: I1004 06:02:48.232795 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:48 crc kubenswrapper[4770]: W1004 06:02:48.269000 4770 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3fb511c_436f_4480_a590_9f7a609ea562.slice/crio-0501f4a168207772f55e9308efdcd59a71f4c1f734b093e0deb6cb8f1e671763 WatchSource:0}: Error finding container 0501f4a168207772f55e9308efdcd59a71f4c1f734b093e0deb6cb8f1e671763: Status 404 returned error can't find the container with id 0501f4a168207772f55e9308efdcd59a71f4c1f734b093e0deb6cb8f1e671763 Oct 04 06:02:48 crc kubenswrapper[4770]: I1004 06:02:48.536045 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/crc-debug-vgq7f" event={"ID":"d3fb511c-436f-4480-a590-9f7a609ea562","Type":"ContainerStarted","Data":"ac8a960fadb00a73be84bb8eab8b63364633c07967b1fe1b754baa919fcbe892"} Oct 04 06:02:48 crc kubenswrapper[4770]: I1004 06:02:48.536132 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/crc-debug-vgq7f" event={"ID":"d3fb511c-436f-4480-a590-9f7a609ea562","Type":"ContainerStarted","Data":"0501f4a168207772f55e9308efdcd59a71f4c1f734b093e0deb6cb8f1e671763"} Oct 04 06:02:49 crc kubenswrapper[4770]: I1004 06:02:49.569339 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c885t/crc-debug-vgq7f" podStartSLOduration=2.569318219 podStartE2EDuration="2.569318219s" podCreationTimestamp="2025-10-04 06:02:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 06:02:49.560182546 +0000 UTC m=+10780.852192268" watchObservedRunningTime="2025-10-04 06:02:49.569318219 +0000 UTC m=+10780.861327961" Oct 04 06:02:50 crc kubenswrapper[4770]: I1004 06:02:50.559515 4770 generic.go:334] "Generic (PLEG): container finished" podID="d3fb511c-436f-4480-a590-9f7a609ea562" containerID="ac8a960fadb00a73be84bb8eab8b63364633c07967b1fe1b754baa919fcbe892" exitCode=0 Oct 04 06:02:50 crc kubenswrapper[4770]: I1004 06:02:50.559593 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/crc-debug-vgq7f" event={"ID":"d3fb511c-436f-4480-a590-9f7a609ea562","Type":"ContainerDied","Data":"ac8a960fadb00a73be84bb8eab8b63364633c07967b1fe1b754baa919fcbe892"} Oct 04 06:02:51 crc kubenswrapper[4770]: I1004 06:02:51.679925 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:51 crc kubenswrapper[4770]: I1004 06:02:51.769491 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz4qz\" (UniqueName: \"kubernetes.io/projected/d3fb511c-436f-4480-a590-9f7a609ea562-kube-api-access-vz4qz\") pod \"d3fb511c-436f-4480-a590-9f7a609ea562\" (UID: \"d3fb511c-436f-4480-a590-9f7a609ea562\") " Oct 04 06:02:51 crc kubenswrapper[4770]: I1004 06:02:51.769625 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3fb511c-436f-4480-a590-9f7a609ea562-host\") pod \"d3fb511c-436f-4480-a590-9f7a609ea562\" (UID: \"d3fb511c-436f-4480-a590-9f7a609ea562\") " Oct 04 06:02:51 crc kubenswrapper[4770]: I1004 06:02:51.769777 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3fb511c-436f-4480-a590-9f7a609ea562-host" (OuterVolumeSpecName: "host") pod "d3fb511c-436f-4480-a590-9f7a609ea562" (UID: "d3fb511c-436f-4480-a590-9f7a609ea562"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 06:02:51 crc kubenswrapper[4770]: I1004 06:02:51.770363 4770 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3fb511c-436f-4480-a590-9f7a609ea562-host\") on node \"crc\" DevicePath \"\"" Oct 04 06:02:51 crc kubenswrapper[4770]: I1004 06:02:51.776135 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3fb511c-436f-4480-a590-9f7a609ea562-kube-api-access-vz4qz" (OuterVolumeSpecName: "kube-api-access-vz4qz") pod "d3fb511c-436f-4480-a590-9f7a609ea562" (UID: "d3fb511c-436f-4480-a590-9f7a609ea562"). InnerVolumeSpecName "kube-api-access-vz4qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:02:51 crc kubenswrapper[4770]: I1004 06:02:51.873560 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz4qz\" (UniqueName: \"kubernetes.io/projected/d3fb511c-436f-4480-a590-9f7a609ea562-kube-api-access-vz4qz\") on node \"crc\" DevicePath \"\"" Oct 04 06:02:52 crc kubenswrapper[4770]: I1004 06:02:52.583535 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/crc-debug-vgq7f" event={"ID":"d3fb511c-436f-4480-a590-9f7a609ea562","Type":"ContainerDied","Data":"0501f4a168207772f55e9308efdcd59a71f4c1f734b093e0deb6cb8f1e671763"} Oct 04 06:02:52 crc kubenswrapper[4770]: I1004 06:02:52.584142 4770 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0501f4a168207772f55e9308efdcd59a71f4c1f734b093e0deb6cb8f1e671763" Oct 04 06:02:52 crc kubenswrapper[4770]: I1004 06:02:52.583699 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-vgq7f" Oct 04 06:02:56 crc kubenswrapper[4770]: I1004 06:02:56.482958 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:56 crc kubenswrapper[4770]: I1004 06:02:56.536737 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8lt8"] Oct 04 06:02:56 crc kubenswrapper[4770]: I1004 06:02:56.627638 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s8lt8" podUID="8c7910d4-ab13-453e-999e-44dad61a6195" containerName="registry-server" containerID="cri-o://3f7040e2a6da53829b35eb259381991ebafbf365b1d9f9667e7e750708c7b25c" gracePeriod=2 Oct 04 06:02:57 crc kubenswrapper[4770]: I1004 06:02:57.649988 4770 generic.go:334] "Generic (PLEG): container finished" podID="8c7910d4-ab13-453e-999e-44dad61a6195" containerID="3f7040e2a6da53829b35eb259381991ebafbf365b1d9f9667e7e750708c7b25c" exitCode=0 Oct 04 06:02:57 crc kubenswrapper[4770]: I1004 06:02:57.650405 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8lt8" event={"ID":"8c7910d4-ab13-453e-999e-44dad61a6195","Type":"ContainerDied","Data":"3f7040e2a6da53829b35eb259381991ebafbf365b1d9f9667e7e750708c7b25c"} Oct 04 06:02:57 crc kubenswrapper[4770]: I1004 06:02:57.771431 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:57 crc kubenswrapper[4770]: I1004 06:02:57.913404 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-catalog-content\") pod \"8c7910d4-ab13-453e-999e-44dad61a6195\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " Oct 04 06:02:57 crc kubenswrapper[4770]: I1004 06:02:57.913526 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2nxw\" (UniqueName: \"kubernetes.io/projected/8c7910d4-ab13-453e-999e-44dad61a6195-kube-api-access-j2nxw\") pod \"8c7910d4-ab13-453e-999e-44dad61a6195\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " Oct 04 06:02:57 crc kubenswrapper[4770]: I1004 06:02:57.913773 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-utilities\") pod \"8c7910d4-ab13-453e-999e-44dad61a6195\" (UID: \"8c7910d4-ab13-453e-999e-44dad61a6195\") " Oct 04 06:02:57 crc kubenswrapper[4770]: I1004 06:02:57.914621 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-utilities" (OuterVolumeSpecName: "utilities") pod "8c7910d4-ab13-453e-999e-44dad61a6195" (UID: "8c7910d4-ab13-453e-999e-44dad61a6195"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:02:57 crc kubenswrapper[4770]: I1004 06:02:57.919182 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c7910d4-ab13-453e-999e-44dad61a6195-kube-api-access-j2nxw" (OuterVolumeSpecName: "kube-api-access-j2nxw") pod "8c7910d4-ab13-453e-999e-44dad61a6195" (UID: "8c7910d4-ab13-453e-999e-44dad61a6195"). InnerVolumeSpecName "kube-api-access-j2nxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:02:57 crc kubenswrapper[4770]: I1004 06:02:57.928653 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c7910d4-ab13-453e-999e-44dad61a6195" (UID: "8c7910d4-ab13-453e-999e-44dad61a6195"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.015920 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.015958 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c7910d4-ab13-453e-999e-44dad61a6195-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.015973 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2nxw\" (UniqueName: \"kubernetes.io/projected/8c7910d4-ab13-453e-999e-44dad61a6195-kube-api-access-j2nxw\") on node \"crc\" DevicePath \"\"" Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.663118 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8lt8" event={"ID":"8c7910d4-ab13-453e-999e-44dad61a6195","Type":"ContainerDied","Data":"a86024dfac9ba8b3a8bc22d55d7c36eec4bee71b925d5dcae4a5eb4ff7dfa229"} Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.663177 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s8lt8" Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.663187 4770 scope.go:117] "RemoveContainer" containerID="3f7040e2a6da53829b35eb259381991ebafbf365b1d9f9667e7e750708c7b25c" Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.699771 4770 scope.go:117] "RemoveContainer" containerID="e597941bca544b04b1b895a7c69ed7428ccc47c173779641cc4ca74219166568" Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.702170 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8lt8"] Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.721604 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8lt8"] Oct 04 06:02:58 crc kubenswrapper[4770]: I1004 06:02:58.723064 4770 scope.go:117] "RemoveContainer" containerID="17fb6fa16512bf2db9426f3319a6f1b3ed7cbdee1c15d331f489208e47aa7f08" Oct 04 06:02:59 crc kubenswrapper[4770]: I1004 06:02:59.690090 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c7910d4-ab13-453e-999e-44dad61a6195" path="/var/lib/kubelet/pods/8c7910d4-ab13-453e-999e-44dad61a6195/volumes" Oct 04 06:03:00 crc kubenswrapper[4770]: I1004 06:03:00.617246 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c885t/crc-debug-vgq7f"] Oct 04 06:03:00 crc kubenswrapper[4770]: I1004 06:03:00.626389 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c885t/crc-debug-vgq7f"] Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.689824 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3fb511c-436f-4480-a590-9f7a609ea562" path="/var/lib/kubelet/pods/d3fb511c-436f-4480-a590-9f7a609ea562/volumes" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.806263 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c885t/crc-debug-2hgfw"] Oct 04 06:03:01 crc kubenswrapper[4770]: E1004 06:03:01.806682 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7910d4-ab13-453e-999e-44dad61a6195" containerName="extract-content" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.806697 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7910d4-ab13-453e-999e-44dad61a6195" containerName="extract-content" Oct 04 06:03:01 crc kubenswrapper[4770]: E1004 06:03:01.806714 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7910d4-ab13-453e-999e-44dad61a6195" containerName="extract-utilities" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.806721 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7910d4-ab13-453e-999e-44dad61a6195" containerName="extract-utilities" Oct 04 06:03:01 crc kubenswrapper[4770]: E1004 06:03:01.806743 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7910d4-ab13-453e-999e-44dad61a6195" containerName="registry-server" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.806749 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7910d4-ab13-453e-999e-44dad61a6195" containerName="registry-server" Oct 04 06:03:01 crc kubenswrapper[4770]: E1004 06:03:01.806777 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fb511c-436f-4480-a590-9f7a609ea562" containerName="container-00" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.806783 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fb511c-436f-4480-a590-9f7a609ea562" containerName="container-00" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.806981 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fb511c-436f-4480-a590-9f7a609ea562" containerName="container-00" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.806996 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7910d4-ab13-453e-999e-44dad61a6195" containerName="registry-server" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.807869 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.897352 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-host\") pod \"crc-debug-2hgfw\" (UID: \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\") " pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:01 crc kubenswrapper[4770]: I1004 06:03:01.897424 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlflk\" (UniqueName: \"kubernetes.io/projected/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-kube-api-access-rlflk\") pod \"crc-debug-2hgfw\" (UID: \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\") " pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.000350 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-host\") pod \"crc-debug-2hgfw\" (UID: \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\") " pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.000687 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlflk\" (UniqueName: \"kubernetes.io/projected/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-kube-api-access-rlflk\") pod \"crc-debug-2hgfw\" (UID: \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\") " pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.000474 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-host\") pod \"crc-debug-2hgfw\" (UID: \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\") " pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.020751 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlflk\" (UniqueName: \"kubernetes.io/projected/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-kube-api-access-rlflk\") pod \"crc-debug-2hgfw\" (UID: \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\") " pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.127478 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.710332 4770 generic.go:334] "Generic (PLEG): container finished" podID="c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13" containerID="812514e18a4c5608c51468c95f9bba5f7afc18ff0da9a6dd05388fffda07731b" exitCode=0 Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.710413 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/crc-debug-2hgfw" event={"ID":"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13","Type":"ContainerDied","Data":"812514e18a4c5608c51468c95f9bba5f7afc18ff0da9a6dd05388fffda07731b"} Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.710674 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/crc-debug-2hgfw" event={"ID":"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13","Type":"ContainerStarted","Data":"397a238fb08f14cfd702f82aed9030bd90de2fdcada42a7f67e0f34a76b83c32"} Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.755167 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c885t/crc-debug-2hgfw"] Oct 04 06:03:02 crc kubenswrapper[4770]: I1004 06:03:02.768060 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c885t/crc-debug-2hgfw"] Oct 04 06:03:03 crc kubenswrapper[4770]: I1004 06:03:03.829059 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:03 crc kubenswrapper[4770]: I1004 06:03:03.940868 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-host\") pod \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\" (UID: \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\") " Oct 04 06:03:03 crc kubenswrapper[4770]: I1004 06:03:03.941000 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlflk\" (UniqueName: \"kubernetes.io/projected/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-kube-api-access-rlflk\") pod \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\" (UID: \"c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13\") " Oct 04 06:03:03 crc kubenswrapper[4770]: I1004 06:03:03.941037 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-host" (OuterVolumeSpecName: "host") pod "c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13" (UID: "c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 06:03:03 crc kubenswrapper[4770]: I1004 06:03:03.941642 4770 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-host\") on node \"crc\" DevicePath \"\"" Oct 04 06:03:03 crc kubenswrapper[4770]: I1004 06:03:03.949365 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-kube-api-access-rlflk" (OuterVolumeSpecName: "kube-api-access-rlflk") pod "c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13" (UID: "c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13"). InnerVolumeSpecName "kube-api-access-rlflk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:03:04 crc kubenswrapper[4770]: I1004 06:03:04.044369 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlflk\" (UniqueName: \"kubernetes.io/projected/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13-kube-api-access-rlflk\") on node \"crc\" DevicePath \"\"" Oct 04 06:03:04 crc kubenswrapper[4770]: I1004 06:03:04.493024 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj_7d1c550a-15df-4478-adf8-eb2a329fa625/util/0.log" Oct 04 06:03:04 crc kubenswrapper[4770]: I1004 06:03:04.734147 4770 scope.go:117] "RemoveContainer" containerID="812514e18a4c5608c51468c95f9bba5f7afc18ff0da9a6dd05388fffda07731b" Oct 04 06:03:04 crc kubenswrapper[4770]: I1004 06:03:04.734295 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/crc-debug-2hgfw" Oct 04 06:03:04 crc kubenswrapper[4770]: I1004 06:03:04.742922 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj_7d1c550a-15df-4478-adf8-eb2a329fa625/util/0.log" Oct 04 06:03:04 crc kubenswrapper[4770]: I1004 06:03:04.775577 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj_7d1c550a-15df-4478-adf8-eb2a329fa625/pull/0.log" Oct 04 06:03:04 crc kubenswrapper[4770]: I1004 06:03:04.785527 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj_7d1c550a-15df-4478-adf8-eb2a329fa625/pull/0.log" Oct 04 06:03:04 crc kubenswrapper[4770]: I1004 06:03:04.995377 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj_7d1c550a-15df-4478-adf8-eb2a329fa625/util/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.001153 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj_7d1c550a-15df-4478-adf8-eb2a329fa625/pull/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.054389 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_41e75d81c5b007a0511a23d4d830a8cbd82f216de848bd7a5f44e4de6ejj9xj_7d1c550a-15df-4478-adf8-eb2a329fa625/extract/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.174698 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5b974f6766-nq6dg_909aa645-88e5-4e67-835e-34854c2f5a74/kube-rbac-proxy/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.296110 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-m269q_c0e405f3-b246-4741-8580-248db52f995d/kube-rbac-proxy/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.313379 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5b974f6766-nq6dg_909aa645-88e5-4e67-835e-34854c2f5a74/manager/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.494066 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-m269q_c0e405f3-b246-4741-8580-248db52f995d/manager/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.527290 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-qj4zt_7a0cc679-8960-4bae-9da9-f790f949b5bf/kube-rbac-proxy/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.620825 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-qj4zt_7a0cc679-8960-4bae-9da9-f790f949b5bf/manager/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.686731 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13" path="/var/lib/kubelet/pods/c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13/volumes" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.738139 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-698456cdc6-dzkch_e26cfeef-31d4-452e-bbf6-f54ab0d3b331/kube-rbac-proxy/0.log" Oct 04 06:03:05 crc kubenswrapper[4770]: I1004 06:03:05.858991 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-698456cdc6-dzkch_e26cfeef-31d4-452e-bbf6-f54ab0d3b331/manager/0.log" Oct 04 06:03:06 crc kubenswrapper[4770]: I1004 06:03:06.209962 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5c497dbdb-bcfrt_aab021e2-b955-4efa-8668-1444f506ca5d/kube-rbac-proxy/0.log" Oct 04 06:03:06 crc kubenswrapper[4770]: I1004 06:03:06.340248 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5c497dbdb-bcfrt_aab021e2-b955-4efa-8668-1444f506ca5d/manager/0.log" Oct 04 06:03:06 crc kubenswrapper[4770]: I1004 06:03:06.386375 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6675647785-zkw64_d35d4909-8960-4d33-9235-9692bf3a4edd/kube-rbac-proxy/0.log" Oct 04 06:03:06 crc kubenswrapper[4770]: I1004 06:03:06.452207 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6675647785-zkw64_d35d4909-8960-4d33-9235-9692bf3a4edd/manager/0.log" Oct 04 06:03:06 crc kubenswrapper[4770]: I1004 06:03:06.586859 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-9tpl9_1fbc59ec-b29e-4843-b34c-43335a590c11/kube-rbac-proxy/0.log" Oct 04 06:03:06 crc kubenswrapper[4770]: I1004 06:03:06.860094 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f5894c49f-56p4n_72f16f95-3791-461e-8d91-d89005030236/kube-rbac-proxy/0.log" Oct 04 06:03:06 crc kubenswrapper[4770]: I1004 06:03:06.895683 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f5894c49f-56p4n_72f16f95-3791-461e-8d91-d89005030236/manager/0.log" Oct 04 06:03:06 crc kubenswrapper[4770]: I1004 06:03:06.910774 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-9tpl9_1fbc59ec-b29e-4843-b34c-43335a590c11/manager/0.log" Oct 04 06:03:07 crc kubenswrapper[4770]: I1004 06:03:07.101381 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-57c9cdcf57-95452_c1bba079-ef51-40bb-b92d-b03a6e31d457/kube-rbac-proxy/0.log" Oct 04 06:03:07 crc kubenswrapper[4770]: I1004 06:03:07.236578 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-57c9cdcf57-95452_c1bba079-ef51-40bb-b92d-b03a6e31d457/manager/0.log" Oct 04 06:03:07 crc kubenswrapper[4770]: I1004 06:03:07.319517 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-rwskr_279505bc-0485-4aa9-87ce-930272978c29/kube-rbac-proxy/0.log" Oct 04 06:03:07 crc kubenswrapper[4770]: I1004 06:03:07.450946 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-rwskr_279505bc-0485-4aa9-87ce-930272978c29/manager/0.log" Oct 04 06:03:07 crc kubenswrapper[4770]: I1004 06:03:07.522795 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-67q4w_09622ed2-b879-4add-9df7-0fb77dfa85cb/kube-rbac-proxy/0.log" Oct 04 06:03:07 crc kubenswrapper[4770]: I1004 06:03:07.602930 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-67q4w_09622ed2-b879-4add-9df7-0fb77dfa85cb/manager/0.log" Oct 04 06:03:07 crc kubenswrapper[4770]: I1004 06:03:07.762080 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-rn8fr_acfbbcdb-767c-4a41-924e-dc60bbd3a5ae/kube-rbac-proxy/0.log" Oct 04 06:03:07 crc kubenswrapper[4770]: I1004 06:03:07.848184 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-rn8fr_acfbbcdb-767c-4a41-924e-dc60bbd3a5ae/manager/0.log" Oct 04 06:03:07 crc kubenswrapper[4770]: I1004 06:03:07.947035 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-sn2wq_fcfac3b9-2e3b-453e-b434-42159c1f92a9/kube-rbac-proxy/0.log" Oct 04 06:03:08 crc kubenswrapper[4770]: I1004 06:03:08.104945 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-fkh9b_c0126071-94d2-4c21-ac54-ea6de14946ef/kube-rbac-proxy/0.log" Oct 04 06:03:08 crc kubenswrapper[4770]: I1004 06:03:08.229612 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-fkh9b_c0126071-94d2-4c21-ac54-ea6de14946ef/manager/0.log" Oct 04 06:03:08 crc kubenswrapper[4770]: I1004 06:03:08.339922 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-sn2wq_fcfac3b9-2e3b-453e-b434-42159c1f92a9/manager/0.log" Oct 04 06:03:08 crc kubenswrapper[4770]: I1004 06:03:08.372224 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6_6231ce27-f748-4e48-ac46-60b1d336e153/kube-rbac-proxy/0.log" Oct 04 06:03:08 crc kubenswrapper[4770]: I1004 06:03:08.534434 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-66cc85b5d5dgxp6_6231ce27-f748-4e48-ac46-60b1d336e153/manager/0.log" Oct 04 06:03:08 crc kubenswrapper[4770]: I1004 06:03:08.571721 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7996577b9d-7kzr4_ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c/kube-rbac-proxy/0.log" Oct 04 06:03:08 crc kubenswrapper[4770]: I1004 06:03:08.761679 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cd5c57ff8-fdfll_c2be3a13-80bd-4d80-9f19-35483ae16fc4/kube-rbac-proxy/0.log" Oct 04 06:03:08 crc kubenswrapper[4770]: I1004 06:03:08.947253 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cd5c57ff8-fdfll_c2be3a13-80bd-4d80-9f19-35483ae16fc4/operator/0.log" Oct 04 06:03:09 crc kubenswrapper[4770]: I1004 06:03:09.099228 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-c968bb45-bd27z_2b8720ec-1142-490f-9a6e-9c94e08db137/kube-rbac-proxy/0.log" Oct 04 06:03:09 crc kubenswrapper[4770]: I1004 06:03:09.225184 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-qgzdn_9751f79d-b495-4ea5-97ca-a8b4fe50d7ca/registry-server/0.log" Oct 04 06:03:09 crc kubenswrapper[4770]: I1004 06:03:09.296115 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-c968bb45-bd27z_2b8720ec-1142-490f-9a6e-9c94e08db137/manager/0.log" Oct 04 06:03:09 crc kubenswrapper[4770]: I1004 06:03:09.321243 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-x8ljn_718d82f1-fa64-4a9d-8780-318b056f44a0/kube-rbac-proxy/0.log" Oct 04 06:03:09 crc kubenswrapper[4770]: I1004 06:03:09.455411 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-x8ljn_718d82f1-fa64-4a9d-8780-318b056f44a0/manager/0.log" Oct 04 06:03:09 crc kubenswrapper[4770]: I1004 06:03:09.547142 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-kwmvv_48592971-02b8-495c-8d9e-80956362e01e/operator/0.log" Oct 04 06:03:09 crc kubenswrapper[4770]: I1004 06:03:09.710767 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-8d8z6_34a19c5f-604f-4225-be1d-c66d51b0919d/kube-rbac-proxy/0.log" Oct 04 06:03:09 crc kubenswrapper[4770]: I1004 06:03:09.878803 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-26xxz_48b4bf81-4d46-4b8c-8a98-2112cb67104c/kube-rbac-proxy/0.log" Oct 04 06:03:09 crc kubenswrapper[4770]: I1004 06:03:09.983454 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-8d8z6_34a19c5f-604f-4225-be1d-c66d51b0919d/manager/0.log" Oct 04 06:03:10 crc kubenswrapper[4770]: I1004 06:03:10.117717 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-mhcm4_64c06e35-b80f-4889-a680-36c627c99a06/kube-rbac-proxy/0.log" Oct 04 06:03:10 crc kubenswrapper[4770]: I1004 06:03:10.261803 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-mhcm4_64c06e35-b80f-4889-a680-36c627c99a06/manager/0.log" Oct 04 06:03:10 crc kubenswrapper[4770]: I1004 06:03:10.268396 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-26xxz_48b4bf81-4d46-4b8c-8a98-2112cb67104c/manager/0.log" Oct 04 06:03:10 crc kubenswrapper[4770]: I1004 06:03:10.362608 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-l5k6p_059b3d3c-de5a-490e-94ff-f15b249de3b3/kube-rbac-proxy/0.log" Oct 04 06:03:10 crc kubenswrapper[4770]: I1004 06:03:10.445587 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-l5k6p_059b3d3c-de5a-490e-94ff-f15b249de3b3/manager/0.log" Oct 04 06:03:11 crc kubenswrapper[4770]: I1004 06:03:11.294656 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7996577b9d-7kzr4_ead820e4-bb4f-4aaa-a35c-fd3f1ac3460c/manager/0.log" Oct 04 06:03:26 crc kubenswrapper[4770]: I1004 06:03:26.879783 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-8srq6_056a8276-f8ee-45f5-a145-63db980dc0c4/control-plane-machine-set-operator/0.log" Oct 04 06:03:27 crc kubenswrapper[4770]: I1004 06:03:27.036165 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fbjnf_0e1a8be1-36ca-4579-9678-13ba76cec6b1/kube-rbac-proxy/0.log" Oct 04 06:03:27 crc kubenswrapper[4770]: I1004 06:03:27.050155 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fbjnf_0e1a8be1-36ca-4579-9678-13ba76cec6b1/machine-api-operator/0.log" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.765688 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t9k24"] Oct 04 06:03:34 crc kubenswrapper[4770]: E1004 06:03:34.766834 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13" containerName="container-00" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.766853 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13" containerName="container-00" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.767188 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4a9dafc-36b4-4c0f-88e7-86df6e8c2f13" containerName="container-00" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.769240 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.815248 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t9k24"] Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.844788 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-catalog-content\") pod \"community-operators-t9k24\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.844863 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx4jf\" (UniqueName: \"kubernetes.io/projected/8bf19c4e-2ea2-427b-83ad-5b266f009c26-kube-api-access-lx4jf\") pod \"community-operators-t9k24\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.845131 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-utilities\") pod \"community-operators-t9k24\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.946858 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-utilities\") pod \"community-operators-t9k24\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.946942 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-catalog-content\") pod \"community-operators-t9k24\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.946971 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx4jf\" (UniqueName: \"kubernetes.io/projected/8bf19c4e-2ea2-427b-83ad-5b266f009c26-kube-api-access-lx4jf\") pod \"community-operators-t9k24\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.947496 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-utilities\") pod \"community-operators-t9k24\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.947556 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-catalog-content\") pod \"community-operators-t9k24\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:34 crc kubenswrapper[4770]: I1004 06:03:34.966284 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx4jf\" (UniqueName: \"kubernetes.io/projected/8bf19c4e-2ea2-427b-83ad-5b266f009c26-kube-api-access-lx4jf\") pod \"community-operators-t9k24\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:35 crc kubenswrapper[4770]: I1004 06:03:35.134830 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:35 crc kubenswrapper[4770]: I1004 06:03:35.750056 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t9k24"] Oct 04 06:03:36 crc kubenswrapper[4770]: I1004 06:03:36.062825 4770 generic.go:334] "Generic (PLEG): container finished" podID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerID="baf2c89928adfdbbc430078825775ec52a143e48622eb2fd7e2909be9781909f" exitCode=0 Oct 04 06:03:36 crc kubenswrapper[4770]: I1004 06:03:36.063227 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9k24" event={"ID":"8bf19c4e-2ea2-427b-83ad-5b266f009c26","Type":"ContainerDied","Data":"baf2c89928adfdbbc430078825775ec52a143e48622eb2fd7e2909be9781909f"} Oct 04 06:03:36 crc kubenswrapper[4770]: I1004 06:03:36.063261 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9k24" event={"ID":"8bf19c4e-2ea2-427b-83ad-5b266f009c26","Type":"ContainerStarted","Data":"12b1cdf048aa24836ddd7a9479193aedb75d2ba46f1db83e00c915b1d3be71b9"} Oct 04 06:03:36 crc kubenswrapper[4770]: I1004 06:03:36.066890 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 06:03:37 crc kubenswrapper[4770]: I1004 06:03:37.078288 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9k24" event={"ID":"8bf19c4e-2ea2-427b-83ad-5b266f009c26","Type":"ContainerStarted","Data":"005488fdde5f213bb74c58a5b60d5130425eabd19604bc0c922d0874c18f4dfa"} Oct 04 06:03:38 crc kubenswrapper[4770]: I1004 06:03:38.096613 4770 generic.go:334] "Generic (PLEG): container finished" podID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerID="005488fdde5f213bb74c58a5b60d5130425eabd19604bc0c922d0874c18f4dfa" exitCode=0 Oct 04 06:03:38 crc kubenswrapper[4770]: I1004 06:03:38.096928 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9k24" event={"ID":"8bf19c4e-2ea2-427b-83ad-5b266f009c26","Type":"ContainerDied","Data":"005488fdde5f213bb74c58a5b60d5130425eabd19604bc0c922d0874c18f4dfa"} Oct 04 06:03:39 crc kubenswrapper[4770]: I1004 06:03:39.111216 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9k24" event={"ID":"8bf19c4e-2ea2-427b-83ad-5b266f009c26","Type":"ContainerStarted","Data":"dc583745936731b0bc285aae852d4436d48ce26f393dc17cd4c2d9669b2a0755"} Oct 04 06:03:39 crc kubenswrapper[4770]: I1004 06:03:39.137480 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t9k24" podStartSLOduration=2.517896582 podStartE2EDuration="5.13745717s" podCreationTimestamp="2025-10-04 06:03:34 +0000 UTC" firstStartedPulling="2025-10-04 06:03:36.066571144 +0000 UTC m=+10827.358580846" lastFinishedPulling="2025-10-04 06:03:38.686131722 +0000 UTC m=+10829.978141434" observedRunningTime="2025-10-04 06:03:39.127162277 +0000 UTC m=+10830.419172019" watchObservedRunningTime="2025-10-04 06:03:39.13745717 +0000 UTC m=+10830.429466882" Oct 04 06:03:39 crc kubenswrapper[4770]: I1004 06:03:39.635677 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-gwpl4_06c54c3d-d425-46b6-aedb-46606d7e9c0d/cert-manager-controller/0.log" Oct 04 06:03:39 crc kubenswrapper[4770]: I1004 06:03:39.919326 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-fwxql_ed3e720d-90d9-4f0b-8b40-edf9eda88615/cert-manager-webhook/0.log" Oct 04 06:03:39 crc kubenswrapper[4770]: I1004 06:03:39.942999 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-298sm_26947e76-92b5-49ad-976f-fb19b0426a3c/cert-manager-cainjector/0.log" Oct 04 06:03:45 crc kubenswrapper[4770]: I1004 06:03:45.135681 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:45 crc kubenswrapper[4770]: I1004 06:03:45.136331 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:45 crc kubenswrapper[4770]: I1004 06:03:45.188040 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:45 crc kubenswrapper[4770]: I1004 06:03:45.237843 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:45 crc kubenswrapper[4770]: I1004 06:03:45.426928 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t9k24"] Oct 04 06:03:47 crc kubenswrapper[4770]: I1004 06:03:47.204886 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t9k24" podUID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerName="registry-server" containerID="cri-o://dc583745936731b0bc285aae852d4436d48ce26f393dc17cd4c2d9669b2a0755" gracePeriod=2 Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.217211 4770 generic.go:334] "Generic (PLEG): container finished" podID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerID="dc583745936731b0bc285aae852d4436d48ce26f393dc17cd4c2d9669b2a0755" exitCode=0 Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.217284 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9k24" event={"ID":"8bf19c4e-2ea2-427b-83ad-5b266f009c26","Type":"ContainerDied","Data":"dc583745936731b0bc285aae852d4436d48ce26f393dc17cd4c2d9669b2a0755"} Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.648442 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.696835 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-catalog-content\") pod \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.696972 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-utilities\") pod \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.697094 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lx4jf\" (UniqueName: \"kubernetes.io/projected/8bf19c4e-2ea2-427b-83ad-5b266f009c26-kube-api-access-lx4jf\") pod \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\" (UID: \"8bf19c4e-2ea2-427b-83ad-5b266f009c26\") " Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.698065 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-utilities" (OuterVolumeSpecName: "utilities") pod "8bf19c4e-2ea2-427b-83ad-5b266f009c26" (UID: "8bf19c4e-2ea2-427b-83ad-5b266f009c26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.708890 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf19c4e-2ea2-427b-83ad-5b266f009c26-kube-api-access-lx4jf" (OuterVolumeSpecName: "kube-api-access-lx4jf") pod "8bf19c4e-2ea2-427b-83ad-5b266f009c26" (UID: "8bf19c4e-2ea2-427b-83ad-5b266f009c26"). InnerVolumeSpecName "kube-api-access-lx4jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.755414 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bf19c4e-2ea2-427b-83ad-5b266f009c26" (UID: "8bf19c4e-2ea2-427b-83ad-5b266f009c26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.799700 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.799747 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lx4jf\" (UniqueName: \"kubernetes.io/projected/8bf19c4e-2ea2-427b-83ad-5b266f009c26-kube-api-access-lx4jf\") on node \"crc\" DevicePath \"\"" Oct 04 06:03:48 crc kubenswrapper[4770]: I1004 06:03:48.799763 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf19c4e-2ea2-427b-83ad-5b266f009c26-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 06:03:49 crc kubenswrapper[4770]: I1004 06:03:49.231068 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9k24" event={"ID":"8bf19c4e-2ea2-427b-83ad-5b266f009c26","Type":"ContainerDied","Data":"12b1cdf048aa24836ddd7a9479193aedb75d2ba46f1db83e00c915b1d3be71b9"} Oct 04 06:03:49 crc kubenswrapper[4770]: I1004 06:03:49.231157 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9k24" Oct 04 06:03:49 crc kubenswrapper[4770]: I1004 06:03:49.231390 4770 scope.go:117] "RemoveContainer" containerID="dc583745936731b0bc285aae852d4436d48ce26f393dc17cd4c2d9669b2a0755" Oct 04 06:03:49 crc kubenswrapper[4770]: I1004 06:03:49.260833 4770 scope.go:117] "RemoveContainer" containerID="005488fdde5f213bb74c58a5b60d5130425eabd19604bc0c922d0874c18f4dfa" Oct 04 06:03:49 crc kubenswrapper[4770]: I1004 06:03:49.268460 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t9k24"] Oct 04 06:03:49 crc kubenswrapper[4770]: I1004 06:03:49.284176 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t9k24"] Oct 04 06:03:49 crc kubenswrapper[4770]: I1004 06:03:49.287957 4770 scope.go:117] "RemoveContainer" containerID="baf2c89928adfdbbc430078825775ec52a143e48622eb2fd7e2909be9781909f" Oct 04 06:03:49 crc kubenswrapper[4770]: I1004 06:03:49.692155 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" path="/var/lib/kubelet/pods/8bf19c4e-2ea2-427b-83ad-5b266f009c26/volumes" Oct 04 06:03:51 crc kubenswrapper[4770]: I1004 06:03:51.832711 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-tvcgh_ddf1a09e-3019-4a04-b64d-288c8b8cc7d8/nmstate-console-plugin/0.log" Oct 04 06:03:52 crc kubenswrapper[4770]: I1004 06:03:52.052148 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8fc4j_e2cb1036-dcd6-467d-a173-a48732ca9da4/kube-rbac-proxy/0.log" Oct 04 06:03:52 crc kubenswrapper[4770]: I1004 06:03:52.052359 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-ghqhj_6ec13f6e-cab1-4429-a048-f473cbddcc8f/nmstate-handler/0.log" Oct 04 06:03:52 crc kubenswrapper[4770]: I1004 06:03:52.074694 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8fc4j_e2cb1036-dcd6-467d-a173-a48732ca9da4/nmstate-metrics/0.log" Oct 04 06:03:52 crc kubenswrapper[4770]: I1004 06:03:52.309174 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-94wmz_2ecd7b0c-5e2d-4ed6-b4aa-cb7023d7fcbb/nmstate-operator/0.log" Oct 04 06:03:52 crc kubenswrapper[4770]: I1004 06:03:52.431601 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-4jxww_d0610d63-ebfc-43a2-9b50-3ca6cee73d14/nmstate-webhook/0.log" Oct 04 06:04:01 crc kubenswrapper[4770]: I1004 06:04:01.796094 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:04:01 crc kubenswrapper[4770]: I1004 06:04:01.796809 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:04:06 crc kubenswrapper[4770]: I1004 06:04:06.566926 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-5vpbh_6c6436ad-4593-41ac-99e1-81e41edcaf00/kube-rbac-proxy/0.log" Oct 04 06:04:06 crc kubenswrapper[4770]: I1004 06:04:06.818886 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-frr-files/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.019214 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-reloader/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.053208 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-5vpbh_6c6436ad-4593-41ac-99e1-81e41edcaf00/controller/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.056456 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-metrics/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.096951 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-frr-files/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.233482 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-reloader/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.412496 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-metrics/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.471277 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-reloader/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.489486 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-metrics/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.513847 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-frr-files/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.708247 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-metrics/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.713595 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-frr-files/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.739341 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/cp-reloader/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.841170 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/controller/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.948402 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/kube-rbac-proxy/0.log" Oct 04 06:04:07 crc kubenswrapper[4770]: I1004 06:04:07.950647 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/frr-metrics/0.log" Oct 04 06:04:08 crc kubenswrapper[4770]: I1004 06:04:08.268587 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/kube-rbac-proxy-frr/0.log" Oct 04 06:04:08 crc kubenswrapper[4770]: I1004 06:04:08.341015 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/reloader/0.log" Oct 04 06:04:08 crc kubenswrapper[4770]: I1004 06:04:08.531414 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-rp264_611c30a9-d9f8-4edd-a69e-07b6eda06a0e/frr-k8s-webhook-server/0.log" Oct 04 06:04:08 crc kubenswrapper[4770]: I1004 06:04:08.701146 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-c8bf5ddcb-9s6db_e75166b5-2b64-4717-9805-50d03fee9a78/manager/0.log" Oct 04 06:04:08 crc kubenswrapper[4770]: I1004 06:04:08.843698 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-659bb64868-tcb2b_e5cd2a2a-3443-4865-af85-d2da6ca9900a/webhook-server/0.log" Oct 04 06:04:09 crc kubenswrapper[4770]: I1004 06:04:09.140699 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4xpdt_1ee98b5e-b3c0-4c5e-a99c-332d513bc539/kube-rbac-proxy/0.log" Oct 04 06:04:10 crc kubenswrapper[4770]: I1004 06:04:10.558458 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4xpdt_1ee98b5e-b3c0-4c5e-a99c-332d513bc539/speaker/0.log" Oct 04 06:04:11 crc kubenswrapper[4770]: I1004 06:04:11.738067 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-crfm8_78a45309-f737-43ad-ab27-8a0f747536a5/frr/0.log" Oct 04 06:04:23 crc kubenswrapper[4770]: I1004 06:04:23.399830 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj_714e499d-6a91-4a5a-b96f-4a56413bc7eb/util/0.log" Oct 04 06:04:23 crc kubenswrapper[4770]: I1004 06:04:23.655779 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj_714e499d-6a91-4a5a-b96f-4a56413bc7eb/util/0.log" Oct 04 06:04:23 crc kubenswrapper[4770]: I1004 06:04:23.694757 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj_714e499d-6a91-4a5a-b96f-4a56413bc7eb/pull/0.log" Oct 04 06:04:23 crc kubenswrapper[4770]: I1004 06:04:23.782734 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj_714e499d-6a91-4a5a-b96f-4a56413bc7eb/pull/0.log" Oct 04 06:04:23 crc kubenswrapper[4770]: I1004 06:04:23.973616 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj_714e499d-6a91-4a5a-b96f-4a56413bc7eb/extract/0.log" Oct 04 06:04:23 crc kubenswrapper[4770]: I1004 06:04:23.975159 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj_714e499d-6a91-4a5a-b96f-4a56413bc7eb/util/0.log" Oct 04 06:04:24 crc kubenswrapper[4770]: I1004 06:04:24.022602 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rr6nj_714e499d-6a91-4a5a-b96f-4a56413bc7eb/pull/0.log" Oct 04 06:04:24 crc kubenswrapper[4770]: I1004 06:04:24.241131 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7_7d6f0476-54a4-43f3-8fe8-737fd8c4dc17/util/0.log" Oct 04 06:04:24 crc kubenswrapper[4770]: I1004 06:04:24.410953 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7_7d6f0476-54a4-43f3-8fe8-737fd8c4dc17/pull/0.log" Oct 04 06:04:24 crc kubenswrapper[4770]: I1004 06:04:24.434890 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7_7d6f0476-54a4-43f3-8fe8-737fd8c4dc17/pull/0.log" Oct 04 06:04:24 crc kubenswrapper[4770]: I1004 06:04:24.441059 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7_7d6f0476-54a4-43f3-8fe8-737fd8c4dc17/util/0.log" Oct 04 06:04:24 crc kubenswrapper[4770]: I1004 06:04:24.684632 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7_7d6f0476-54a4-43f3-8fe8-737fd8c4dc17/util/0.log" Oct 04 06:04:24 crc kubenswrapper[4770]: I1004 06:04:24.731696 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7_7d6f0476-54a4-43f3-8fe8-737fd8c4dc17/pull/0.log" Oct 04 06:04:24 crc kubenswrapper[4770]: I1004 06:04:24.769407 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2r4rz7_7d6f0476-54a4-43f3-8fe8-737fd8c4dc17/extract/0.log" Oct 04 06:04:24 crc kubenswrapper[4770]: I1004 06:04:24.908226 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr_4129621b-ab45-4ed3-b43f-590caee83445/util/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.168971 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr_4129621b-ab45-4ed3-b43f-590caee83445/util/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.180570 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr_4129621b-ab45-4ed3-b43f-590caee83445/pull/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.211868 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr_4129621b-ab45-4ed3-b43f-590caee83445/pull/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.348854 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr_4129621b-ab45-4ed3-b43f-590caee83445/util/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.390495 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr_4129621b-ab45-4ed3-b43f-590caee83445/pull/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.393744 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d798hr_4129621b-ab45-4ed3-b43f-590caee83445/extract/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.560546 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5h6qk_10bb88a1-02bc-4344-9f59-1e4794a719a0/extract-utilities/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.746115 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5h6qk_10bb88a1-02bc-4344-9f59-1e4794a719a0/extract-content/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.748469 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5h6qk_10bb88a1-02bc-4344-9f59-1e4794a719a0/extract-content/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.754870 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5h6qk_10bb88a1-02bc-4344-9f59-1e4794a719a0/extract-utilities/0.log" Oct 04 06:04:25 crc kubenswrapper[4770]: I1004 06:04:25.991985 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5h6qk_10bb88a1-02bc-4344-9f59-1e4794a719a0/extract-utilities/0.log" Oct 04 06:04:26 crc kubenswrapper[4770]: I1004 06:04:26.021002 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5h6qk_10bb88a1-02bc-4344-9f59-1e4794a719a0/extract-content/0.log" Oct 04 06:04:26 crc kubenswrapper[4770]: I1004 06:04:26.275192 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bf524_39a3f00b-87d7-4580-a778-2bc608aef9ce/extract-utilities/0.log" Oct 04 06:04:26 crc kubenswrapper[4770]: I1004 06:04:26.549660 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bf524_39a3f00b-87d7-4580-a778-2bc608aef9ce/extract-content/0.log" Oct 04 06:04:26 crc kubenswrapper[4770]: I1004 06:04:26.556929 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bf524_39a3f00b-87d7-4580-a778-2bc608aef9ce/extract-utilities/0.log" Oct 04 06:04:26 crc kubenswrapper[4770]: I1004 06:04:26.571770 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bf524_39a3f00b-87d7-4580-a778-2bc608aef9ce/extract-content/0.log" Oct 04 06:04:27 crc kubenswrapper[4770]: I1004 06:04:27.050777 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bf524_39a3f00b-87d7-4580-a778-2bc608aef9ce/extract-utilities/0.log" Oct 04 06:04:27 crc kubenswrapper[4770]: I1004 06:04:27.100881 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bf524_39a3f00b-87d7-4580-a778-2bc608aef9ce/extract-content/0.log" Oct 04 06:04:27 crc kubenswrapper[4770]: I1004 06:04:27.484121 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s_5e4248c4-b986-498f-b700-f56c4eb3e29d/util/0.log" Oct 04 06:04:27 crc kubenswrapper[4770]: I1004 06:04:27.591865 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s_5e4248c4-b986-498f-b700-f56c4eb3e29d/util/0.log" Oct 04 06:04:27 crc kubenswrapper[4770]: I1004 06:04:27.647890 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s_5e4248c4-b986-498f-b700-f56c4eb3e29d/pull/0.log" Oct 04 06:04:27 crc kubenswrapper[4770]: I1004 06:04:27.768666 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bf524_39a3f00b-87d7-4580-a778-2bc608aef9ce/registry-server/0.log" Oct 04 06:04:27 crc kubenswrapper[4770]: I1004 06:04:27.802890 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s_5e4248c4-b986-498f-b700-f56c4eb3e29d/pull/0.log" Oct 04 06:04:27 crc kubenswrapper[4770]: I1004 06:04:27.803838 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5h6qk_10bb88a1-02bc-4344-9f59-1e4794a719a0/registry-server/0.log" Oct 04 06:04:27 crc kubenswrapper[4770]: I1004 06:04:27.983053 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s_5e4248c4-b986-498f-b700-f56c4eb3e29d/pull/0.log" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.003056 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s_5e4248c4-b986-498f-b700-f56c4eb3e29d/util/0.log" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.091573 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-46c6n_080999d9-34f8-445c-b305-e41097107afc/marketplace-operator/0.log" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.112211 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7pp8s_5e4248c4-b986-498f-b700-f56c4eb3e29d/extract/0.log" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.268535 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ppnbx_14b26c94-8153-4918-afe0-098a8e16a725/extract-utilities/0.log" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.329231 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sc6qt"] Oct 04 06:04:28 crc kubenswrapper[4770]: E1004 06:04:28.329821 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerName="registry-server" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.329886 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerName="registry-server" Oct 04 06:04:28 crc kubenswrapper[4770]: E1004 06:04:28.329936 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerName="extract-utilities" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.329945 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerName="extract-utilities" Oct 04 06:04:28 crc kubenswrapper[4770]: E1004 06:04:28.329977 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerName="extract-content" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.329987 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerName="extract-content" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.330259 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf19c4e-2ea2-427b-83ad-5b266f009c26" containerName="registry-server" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.332613 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.345760 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sc6qt"] Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.417716 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-catalog-content\") pod \"certified-operators-sc6qt\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.417883 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-utilities\") pod \"certified-operators-sc6qt\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.417940 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-974jv\" (UniqueName: \"kubernetes.io/projected/85b4f207-7782-4e22-a944-13da46eaf407-kube-api-access-974jv\") pod \"certified-operators-sc6qt\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.519677 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-utilities\") pod \"certified-operators-sc6qt\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.519828 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-974jv\" (UniqueName: \"kubernetes.io/projected/85b4f207-7782-4e22-a944-13da46eaf407-kube-api-access-974jv\") pod \"certified-operators-sc6qt\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.519940 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-catalog-content\") pod \"certified-operators-sc6qt\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.520206 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-utilities\") pod \"certified-operators-sc6qt\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.520262 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-catalog-content\") pod \"certified-operators-sc6qt\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.543461 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-974jv\" (UniqueName: \"kubernetes.io/projected/85b4f207-7782-4e22-a944-13da46eaf407-kube-api-access-974jv\") pod \"certified-operators-sc6qt\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.552944 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ppnbx_14b26c94-8153-4918-afe0-098a8e16a725/extract-content/0.log" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.581942 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ppnbx_14b26c94-8153-4918-afe0-098a8e16a725/extract-content/0.log" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.612188 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ppnbx_14b26c94-8153-4918-afe0-098a8e16a725/extract-utilities/0.log" Oct 04 06:04:28 crc kubenswrapper[4770]: I1004 06:04:28.661194 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.279037 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ppnbx_14b26c94-8153-4918-afe0-098a8e16a725/extract-content/0.log" Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.314332 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sc6qt"] Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.343420 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ppnbx_14b26c94-8153-4918-afe0-098a8e16a725/extract-utilities/0.log" Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.344501 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvtk8_071fc5ae-cc19-4618-aa7f-53518791bb22/extract-utilities/0.log" Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.568561 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvtk8_071fc5ae-cc19-4618-aa7f-53518791bb22/extract-utilities/0.log" Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.607305 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvtk8_071fc5ae-cc19-4618-aa7f-53518791bb22/extract-content/0.log" Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.625923 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvtk8_071fc5ae-cc19-4618-aa7f-53518791bb22/extract-content/0.log" Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.737486 4770 generic.go:334] "Generic (PLEG): container finished" podID="85b4f207-7782-4e22-a944-13da46eaf407" containerID="31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c" exitCode=0 Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.737539 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sc6qt" event={"ID":"85b4f207-7782-4e22-a944-13da46eaf407","Type":"ContainerDied","Data":"31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c"} Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.737568 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sc6qt" event={"ID":"85b4f207-7782-4e22-a944-13da46eaf407","Type":"ContainerStarted","Data":"0b289f39b8cb749e2a439230d301ae39fca4a9f74bc051f21af0057709152024"} Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.749430 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ppnbx_14b26c94-8153-4918-afe0-098a8e16a725/registry-server/0.log" Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.911356 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvtk8_071fc5ae-cc19-4618-aa7f-53518791bb22/extract-content/0.log" Oct 04 06:04:29 crc kubenswrapper[4770]: I1004 06:04:29.927216 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvtk8_071fc5ae-cc19-4618-aa7f-53518791bb22/extract-utilities/0.log" Oct 04 06:04:30 crc kubenswrapper[4770]: I1004 06:04:30.750780 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sc6qt" event={"ID":"85b4f207-7782-4e22-a944-13da46eaf407","Type":"ContainerStarted","Data":"5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6"} Oct 04 06:04:31 crc kubenswrapper[4770]: I1004 06:04:31.321461 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvtk8_071fc5ae-cc19-4618-aa7f-53518791bb22/registry-server/0.log" Oct 04 06:04:31 crc kubenswrapper[4770]: I1004 06:04:31.795558 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:04:31 crc kubenswrapper[4770]: I1004 06:04:31.795622 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:04:32 crc kubenswrapper[4770]: I1004 06:04:32.772467 4770 generic.go:334] "Generic (PLEG): container finished" podID="85b4f207-7782-4e22-a944-13da46eaf407" containerID="5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6" exitCode=0 Oct 04 06:04:32 crc kubenswrapper[4770]: I1004 06:04:32.772666 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sc6qt" event={"ID":"85b4f207-7782-4e22-a944-13da46eaf407","Type":"ContainerDied","Data":"5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6"} Oct 04 06:04:33 crc kubenswrapper[4770]: I1004 06:04:33.799927 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sc6qt" event={"ID":"85b4f207-7782-4e22-a944-13da46eaf407","Type":"ContainerStarted","Data":"37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42"} Oct 04 06:04:33 crc kubenswrapper[4770]: I1004 06:04:33.820540 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sc6qt" podStartSLOduration=2.329263866 podStartE2EDuration="5.820520158s" podCreationTimestamp="2025-10-04 06:04:28 +0000 UTC" firstStartedPulling="2025-10-04 06:04:29.739930369 +0000 UTC m=+10881.031940081" lastFinishedPulling="2025-10-04 06:04:33.231186661 +0000 UTC m=+10884.523196373" observedRunningTime="2025-10-04 06:04:33.81756608 +0000 UTC m=+10885.109575792" watchObservedRunningTime="2025-10-04 06:04:33.820520158 +0000 UTC m=+10885.112529880" Oct 04 06:04:38 crc kubenswrapper[4770]: I1004 06:04:38.661581 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:38 crc kubenswrapper[4770]: I1004 06:04:38.662202 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:38 crc kubenswrapper[4770]: I1004 06:04:38.722656 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:38 crc kubenswrapper[4770]: I1004 06:04:38.916453 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:38 crc kubenswrapper[4770]: I1004 06:04:38.969795 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sc6qt"] Oct 04 06:04:40 crc kubenswrapper[4770]: I1004 06:04:40.863031 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sc6qt" podUID="85b4f207-7782-4e22-a944-13da46eaf407" containerName="registry-server" containerID="cri-o://37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42" gracePeriod=2 Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.406390 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.452387 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-utilities\") pod \"85b4f207-7782-4e22-a944-13da46eaf407\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.452796 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-974jv\" (UniqueName: \"kubernetes.io/projected/85b4f207-7782-4e22-a944-13da46eaf407-kube-api-access-974jv\") pod \"85b4f207-7782-4e22-a944-13da46eaf407\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.453068 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-catalog-content\") pod \"85b4f207-7782-4e22-a944-13da46eaf407\" (UID: \"85b4f207-7782-4e22-a944-13da46eaf407\") " Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.453615 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-utilities" (OuterVolumeSpecName: "utilities") pod "85b4f207-7782-4e22-a944-13da46eaf407" (UID: "85b4f207-7782-4e22-a944-13da46eaf407"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.460062 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85b4f207-7782-4e22-a944-13da46eaf407-kube-api-access-974jv" (OuterVolumeSpecName: "kube-api-access-974jv") pod "85b4f207-7782-4e22-a944-13da46eaf407" (UID: "85b4f207-7782-4e22-a944-13da46eaf407"). InnerVolumeSpecName "kube-api-access-974jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.521449 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85b4f207-7782-4e22-a944-13da46eaf407" (UID: "85b4f207-7782-4e22-a944-13da46eaf407"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.560067 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.560106 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b4f207-7782-4e22-a944-13da46eaf407-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.560122 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-974jv\" (UniqueName: \"kubernetes.io/projected/85b4f207-7782-4e22-a944-13da46eaf407-kube-api-access-974jv\") on node \"crc\" DevicePath \"\"" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.882911 4770 generic.go:334] "Generic (PLEG): container finished" podID="85b4f207-7782-4e22-a944-13da46eaf407" containerID="37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42" exitCode=0 Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.882954 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sc6qt" event={"ID":"85b4f207-7782-4e22-a944-13da46eaf407","Type":"ContainerDied","Data":"37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42"} Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.882990 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sc6qt" event={"ID":"85b4f207-7782-4e22-a944-13da46eaf407","Type":"ContainerDied","Data":"0b289f39b8cb749e2a439230d301ae39fca4a9f74bc051f21af0057709152024"} Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.883009 4770 scope.go:117] "RemoveContainer" containerID="37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.883025 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sc6qt" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.907055 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sc6qt"] Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.918293 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sc6qt"] Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.918522 4770 scope.go:117] "RemoveContainer" containerID="5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.950753 4770 scope.go:117] "RemoveContainer" containerID="31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.998220 4770 scope.go:117] "RemoveContainer" containerID="37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42" Oct 04 06:04:41 crc kubenswrapper[4770]: E1004 06:04:41.998761 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42\": container with ID starting with 37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42 not found: ID does not exist" containerID="37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.998801 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42"} err="failed to get container status \"37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42\": rpc error: code = NotFound desc = could not find container \"37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42\": container with ID starting with 37f1598075f43b4bac4c31617f6d04931ba57233403c236d971415afd1741b42 not found: ID does not exist" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.998843 4770 scope.go:117] "RemoveContainer" containerID="5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6" Oct 04 06:04:41 crc kubenswrapper[4770]: E1004 06:04:41.999377 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6\": container with ID starting with 5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6 not found: ID does not exist" containerID="5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.999417 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6"} err="failed to get container status \"5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6\": rpc error: code = NotFound desc = could not find container \"5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6\": container with ID starting with 5a02dc586154778fcea8939537113689e67b2f2dbde9127cee6524f3ff25a8d6 not found: ID does not exist" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.999447 4770 scope.go:117] "RemoveContainer" containerID="31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c" Oct 04 06:04:41 crc kubenswrapper[4770]: E1004 06:04:41.999861 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c\": container with ID starting with 31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c not found: ID does not exist" containerID="31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c" Oct 04 06:04:41 crc kubenswrapper[4770]: I1004 06:04:41.999886 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c"} err="failed to get container status \"31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c\": rpc error: code = NotFound desc = could not find container \"31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c\": container with ID starting with 31b847b944f4fc3564b272d8d941a5b82c11102f706d86b789aa31f2429ded2c not found: ID does not exist" Oct 04 06:04:42 crc kubenswrapper[4770]: I1004 06:04:42.693918 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-d5cfv_b55bd1e1-8afd-42a0-b180-c91c3f66fb4d/prometheus-operator/0.log" Oct 04 06:04:42 crc kubenswrapper[4770]: I1004 06:04:42.886428 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-65bd8845d7-2nldv_f4536236-3a5b-4b7c-8150-aa52693e03cb/prometheus-operator-admission-webhook/0.log" Oct 04 06:04:42 crc kubenswrapper[4770]: I1004 06:04:42.962207 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-65bd8845d7-dsfzb_6b3ecc0f-c11f-4572-8303-65ba8f955b03/prometheus-operator-admission-webhook/0.log" Oct 04 06:04:43 crc kubenswrapper[4770]: I1004 06:04:43.133779 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-5nfbm_3e942fae-95d7-4a1c-bd99-d049c23d15cd/operator/0.log" Oct 04 06:04:43 crc kubenswrapper[4770]: I1004 06:04:43.187433 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-qwcjq_54e760b6-1a04-4551-8fc4-d461748805d4/perses-operator/0.log" Oct 04 06:04:43 crc kubenswrapper[4770]: I1004 06:04:43.692604 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85b4f207-7782-4e22-a944-13da46eaf407" path="/var/lib/kubelet/pods/85b4f207-7782-4e22-a944-13da46eaf407/volumes" Oct 04 06:05:01 crc kubenswrapper[4770]: I1004 06:05:01.795592 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:05:01 crc kubenswrapper[4770]: I1004 06:05:01.796266 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:05:01 crc kubenswrapper[4770]: I1004 06:05:01.796324 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 06:05:01 crc kubenswrapper[4770]: I1004 06:05:01.797605 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2281e97bdf52b9438829f5ef88f5250486a7610080a67d0b0aa67a36806cc454"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 06:05:01 crc kubenswrapper[4770]: I1004 06:05:01.797671 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://2281e97bdf52b9438829f5ef88f5250486a7610080a67d0b0aa67a36806cc454" gracePeriod=600 Oct 04 06:05:02 crc kubenswrapper[4770]: I1004 06:05:02.103693 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="2281e97bdf52b9438829f5ef88f5250486a7610080a67d0b0aa67a36806cc454" exitCode=0 Oct 04 06:05:02 crc kubenswrapper[4770]: I1004 06:05:02.103729 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"2281e97bdf52b9438829f5ef88f5250486a7610080a67d0b0aa67a36806cc454"} Oct 04 06:05:02 crc kubenswrapper[4770]: I1004 06:05:02.104106 4770 scope.go:117] "RemoveContainer" containerID="205e2698364f168875399d94ef4d7763524b130c8fdb7f61eabe63d51a4c6444" Oct 04 06:05:03 crc kubenswrapper[4770]: I1004 06:05:03.127556 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerStarted","Data":"ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588"} Oct 04 06:05:33 crc kubenswrapper[4770]: I1004 06:05:33.654904 4770 scope.go:117] "RemoveContainer" containerID="f2384b679082368582479583a5c36a28ab4d7a321803c54c10c1fdb1166afd84" Oct 04 06:07:31 crc kubenswrapper[4770]: I1004 06:07:31.795947 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:07:31 crc kubenswrapper[4770]: I1004 06:07:31.796645 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:07:45 crc kubenswrapper[4770]: I1004 06:07:45.843985 4770 generic.go:334] "Generic (PLEG): container finished" podID="b0ad49de-f48e-451b-9382-e0729a649190" containerID="73f14e7309b5309b7c2490eb028a1eea77e7a587862c7c1a67048ec344d0b3cf" exitCode=0 Oct 04 06:07:45 crc kubenswrapper[4770]: I1004 06:07:45.844043 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c885t/must-gather-tnff4" event={"ID":"b0ad49de-f48e-451b-9382-e0729a649190","Type":"ContainerDied","Data":"73f14e7309b5309b7c2490eb028a1eea77e7a587862c7c1a67048ec344d0b3cf"} Oct 04 06:07:45 crc kubenswrapper[4770]: I1004 06:07:45.845234 4770 scope.go:117] "RemoveContainer" containerID="73f14e7309b5309b7c2490eb028a1eea77e7a587862c7c1a67048ec344d0b3cf" Oct 04 06:07:46 crc kubenswrapper[4770]: I1004 06:07:46.049121 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c885t_must-gather-tnff4_b0ad49de-f48e-451b-9382-e0729a649190/gather/0.log" Oct 04 06:07:55 crc kubenswrapper[4770]: I1004 06:07:55.783736 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c885t/must-gather-tnff4"] Oct 04 06:07:55 crc kubenswrapper[4770]: I1004 06:07:55.784733 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-c885t/must-gather-tnff4" podUID="b0ad49de-f48e-451b-9382-e0729a649190" containerName="copy" containerID="cri-o://afe7316e47faa286fde3ea0d7559f989e5fff996be69c1f89695e896cf2b4509" gracePeriod=2 Oct 04 06:07:55 crc kubenswrapper[4770]: I1004 06:07:55.793402 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c885t/must-gather-tnff4"] Oct 04 06:07:55 crc kubenswrapper[4770]: I1004 06:07:55.978121 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c885t_must-gather-tnff4_b0ad49de-f48e-451b-9382-e0729a649190/copy/0.log" Oct 04 06:07:55 crc kubenswrapper[4770]: I1004 06:07:55.987793 4770 generic.go:334] "Generic (PLEG): container finished" podID="b0ad49de-f48e-451b-9382-e0729a649190" containerID="afe7316e47faa286fde3ea0d7559f989e5fff996be69c1f89695e896cf2b4509" exitCode=143 Oct 04 06:07:56 crc kubenswrapper[4770]: I1004 06:07:56.283781 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c885t_must-gather-tnff4_b0ad49de-f48e-451b-9382-e0729a649190/copy/0.log" Oct 04 06:07:56 crc kubenswrapper[4770]: I1004 06:07:56.285637 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 06:07:56 crc kubenswrapper[4770]: I1004 06:07:56.382953 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsfp6\" (UniqueName: \"kubernetes.io/projected/b0ad49de-f48e-451b-9382-e0729a649190-kube-api-access-gsfp6\") pod \"b0ad49de-f48e-451b-9382-e0729a649190\" (UID: \"b0ad49de-f48e-451b-9382-e0729a649190\") " Oct 04 06:07:56 crc kubenswrapper[4770]: I1004 06:07:56.383278 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0ad49de-f48e-451b-9382-e0729a649190-must-gather-output\") pod \"b0ad49de-f48e-451b-9382-e0729a649190\" (UID: \"b0ad49de-f48e-451b-9382-e0729a649190\") " Oct 04 06:07:56 crc kubenswrapper[4770]: I1004 06:07:56.393248 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ad49de-f48e-451b-9382-e0729a649190-kube-api-access-gsfp6" (OuterVolumeSpecName: "kube-api-access-gsfp6") pod "b0ad49de-f48e-451b-9382-e0729a649190" (UID: "b0ad49de-f48e-451b-9382-e0729a649190"). InnerVolumeSpecName "kube-api-access-gsfp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:07:56 crc kubenswrapper[4770]: I1004 06:07:56.486397 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsfp6\" (UniqueName: \"kubernetes.io/projected/b0ad49de-f48e-451b-9382-e0729a649190-kube-api-access-gsfp6\") on node \"crc\" DevicePath \"\"" Oct 04 06:07:56 crc kubenswrapper[4770]: I1004 06:07:56.607403 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ad49de-f48e-451b-9382-e0729a649190-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b0ad49de-f48e-451b-9382-e0729a649190" (UID: "b0ad49de-f48e-451b-9382-e0729a649190"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:07:56 crc kubenswrapper[4770]: I1004 06:07:56.691741 4770 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b0ad49de-f48e-451b-9382-e0729a649190-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 06:07:57 crc kubenswrapper[4770]: I1004 06:07:57.000867 4770 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c885t_must-gather-tnff4_b0ad49de-f48e-451b-9382-e0729a649190/copy/0.log" Oct 04 06:07:57 crc kubenswrapper[4770]: I1004 06:07:57.001589 4770 scope.go:117] "RemoveContainer" containerID="afe7316e47faa286fde3ea0d7559f989e5fff996be69c1f89695e896cf2b4509" Oct 04 06:07:57 crc kubenswrapper[4770]: I1004 06:07:57.001713 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c885t/must-gather-tnff4" Oct 04 06:07:57 crc kubenswrapper[4770]: I1004 06:07:57.022690 4770 scope.go:117] "RemoveContainer" containerID="73f14e7309b5309b7c2490eb028a1eea77e7a587862c7c1a67048ec344d0b3cf" Oct 04 06:07:57 crc kubenswrapper[4770]: I1004 06:07:57.686802 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ad49de-f48e-451b-9382-e0729a649190" path="/var/lib/kubelet/pods/b0ad49de-f48e-451b-9382-e0729a649190/volumes" Oct 04 06:08:01 crc kubenswrapper[4770]: I1004 06:08:01.795721 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:08:01 crc kubenswrapper[4770]: I1004 06:08:01.796383 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:08:31 crc kubenswrapper[4770]: I1004 06:08:31.795971 4770 patch_prober.go:28] interesting pod/machine-config-daemon-h6msx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 06:08:31 crc kubenswrapper[4770]: I1004 06:08:31.796775 4770 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 06:08:31 crc kubenswrapper[4770]: I1004 06:08:31.796829 4770 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" Oct 04 06:08:31 crc kubenswrapper[4770]: I1004 06:08:31.797737 4770 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588"} pod="openshift-machine-config-operator/machine-config-daemon-h6msx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 06:08:31 crc kubenswrapper[4770]: I1004 06:08:31.797796 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerName="machine-config-daemon" containerID="cri-o://ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" gracePeriod=600 Oct 04 06:08:31 crc kubenswrapper[4770]: E1004 06:08:31.935085 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:08:32 crc kubenswrapper[4770]: I1004 06:08:32.444023 4770 generic.go:334] "Generic (PLEG): container finished" podID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" exitCode=0 Oct 04 06:08:32 crc kubenswrapper[4770]: I1004 06:08:32.444042 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" event={"ID":"6bd3be93-1791-4cd1-b3ae-b4032548e93a","Type":"ContainerDied","Data":"ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588"} Oct 04 06:08:32 crc kubenswrapper[4770]: I1004 06:08:32.444403 4770 scope.go:117] "RemoveContainer" containerID="2281e97bdf52b9438829f5ef88f5250486a7610080a67d0b0aa67a36806cc454" Oct 04 06:08:32 crc kubenswrapper[4770]: I1004 06:08:32.445281 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:08:32 crc kubenswrapper[4770]: E1004 06:08:32.445750 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:08:47 crc kubenswrapper[4770]: I1004 06:08:47.673618 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:08:47 crc kubenswrapper[4770]: E1004 06:08:47.674557 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:09:00 crc kubenswrapper[4770]: I1004 06:09:00.674784 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:09:00 crc kubenswrapper[4770]: E1004 06:09:00.676182 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:09:12 crc kubenswrapper[4770]: I1004 06:09:12.674457 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:09:12 crc kubenswrapper[4770]: E1004 06:09:12.675539 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:09:23 crc kubenswrapper[4770]: I1004 06:09:23.677731 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:09:23 crc kubenswrapper[4770]: E1004 06:09:23.679205 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.908089 4770 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9p5z7"] Oct 04 06:09:24 crc kubenswrapper[4770]: E1004 06:09:24.908597 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ad49de-f48e-451b-9382-e0729a649190" containerName="copy" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.908613 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ad49de-f48e-451b-9382-e0729a649190" containerName="copy" Oct 04 06:09:24 crc kubenswrapper[4770]: E1004 06:09:24.908632 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b4f207-7782-4e22-a944-13da46eaf407" containerName="extract-utilities" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.908641 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b4f207-7782-4e22-a944-13da46eaf407" containerName="extract-utilities" Oct 04 06:09:24 crc kubenswrapper[4770]: E1004 06:09:24.908654 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b4f207-7782-4e22-a944-13da46eaf407" containerName="extract-content" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.908661 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b4f207-7782-4e22-a944-13da46eaf407" containerName="extract-content" Oct 04 06:09:24 crc kubenswrapper[4770]: E1004 06:09:24.908676 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b4f207-7782-4e22-a944-13da46eaf407" containerName="registry-server" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.908682 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b4f207-7782-4e22-a944-13da46eaf407" containerName="registry-server" Oct 04 06:09:24 crc kubenswrapper[4770]: E1004 06:09:24.908716 4770 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ad49de-f48e-451b-9382-e0729a649190" containerName="gather" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.908723 4770 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ad49de-f48e-451b-9382-e0729a649190" containerName="gather" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.908999 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="85b4f207-7782-4e22-a944-13da46eaf407" containerName="registry-server" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.909036 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ad49de-f48e-451b-9382-e0729a649190" containerName="copy" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.909067 4770 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ad49de-f48e-451b-9382-e0729a649190" containerName="gather" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.911045 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.923718 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9p5z7"] Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.985523 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-catalog-content\") pod \"redhat-operators-9p5z7\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.985855 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-utilities\") pod \"redhat-operators-9p5z7\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:24 crc kubenswrapper[4770]: I1004 06:09:24.985942 4770 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtvds\" (UniqueName: \"kubernetes.io/projected/c14839de-6eee-42e9-a457-e66027e2d998-kube-api-access-qtvds\") pod \"redhat-operators-9p5z7\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:25 crc kubenswrapper[4770]: I1004 06:09:25.088073 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtvds\" (UniqueName: \"kubernetes.io/projected/c14839de-6eee-42e9-a457-e66027e2d998-kube-api-access-qtvds\") pod \"redhat-operators-9p5z7\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:25 crc kubenswrapper[4770]: I1004 06:09:25.088195 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-catalog-content\") pod \"redhat-operators-9p5z7\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:25 crc kubenswrapper[4770]: I1004 06:09:25.088299 4770 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-utilities\") pod \"redhat-operators-9p5z7\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:25 crc kubenswrapper[4770]: I1004 06:09:25.088609 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-catalog-content\") pod \"redhat-operators-9p5z7\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:25 crc kubenswrapper[4770]: I1004 06:09:25.088633 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-utilities\") pod \"redhat-operators-9p5z7\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:25 crc kubenswrapper[4770]: I1004 06:09:25.115086 4770 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtvds\" (UniqueName: \"kubernetes.io/projected/c14839de-6eee-42e9-a457-e66027e2d998-kube-api-access-qtvds\") pod \"redhat-operators-9p5z7\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:25 crc kubenswrapper[4770]: I1004 06:09:25.238920 4770 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:25 crc kubenswrapper[4770]: I1004 06:09:25.747756 4770 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9p5z7"] Oct 04 06:09:26 crc kubenswrapper[4770]: I1004 06:09:26.036820 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p5z7" event={"ID":"c14839de-6eee-42e9-a457-e66027e2d998","Type":"ContainerStarted","Data":"36fdb0f37a3faf5794384408e7807181c078323b9958e6347860375d9a0fd637"} Oct 04 06:09:27 crc kubenswrapper[4770]: I1004 06:09:27.055873 4770 generic.go:334] "Generic (PLEG): container finished" podID="c14839de-6eee-42e9-a457-e66027e2d998" containerID="76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34" exitCode=0 Oct 04 06:09:27 crc kubenswrapper[4770]: I1004 06:09:27.055914 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p5z7" event={"ID":"c14839de-6eee-42e9-a457-e66027e2d998","Type":"ContainerDied","Data":"76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34"} Oct 04 06:09:27 crc kubenswrapper[4770]: I1004 06:09:27.058227 4770 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 06:09:29 crc kubenswrapper[4770]: I1004 06:09:29.082326 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p5z7" event={"ID":"c14839de-6eee-42e9-a457-e66027e2d998","Type":"ContainerStarted","Data":"7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363"} Oct 04 06:09:32 crc kubenswrapper[4770]: I1004 06:09:32.115973 4770 generic.go:334] "Generic (PLEG): container finished" podID="c14839de-6eee-42e9-a457-e66027e2d998" containerID="7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363" exitCode=0 Oct 04 06:09:32 crc kubenswrapper[4770]: I1004 06:09:32.116078 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p5z7" event={"ID":"c14839de-6eee-42e9-a457-e66027e2d998","Type":"ContainerDied","Data":"7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363"} Oct 04 06:09:33 crc kubenswrapper[4770]: I1004 06:09:33.141544 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p5z7" event={"ID":"c14839de-6eee-42e9-a457-e66027e2d998","Type":"ContainerStarted","Data":"f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c"} Oct 04 06:09:33 crc kubenswrapper[4770]: I1004 06:09:33.175275 4770 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9p5z7" podStartSLOduration=3.702679656 podStartE2EDuration="9.1752436s" podCreationTimestamp="2025-10-04 06:09:24 +0000 UTC" firstStartedPulling="2025-10-04 06:09:27.057983109 +0000 UTC m=+11178.349992821" lastFinishedPulling="2025-10-04 06:09:32.530547053 +0000 UTC m=+11183.822556765" observedRunningTime="2025-10-04 06:09:33.169116011 +0000 UTC m=+11184.461125743" watchObservedRunningTime="2025-10-04 06:09:33.1752436 +0000 UTC m=+11184.467253312" Oct 04 06:09:33 crc kubenswrapper[4770]: I1004 06:09:33.801580 4770 scope.go:117] "RemoveContainer" containerID="ac8a960fadb00a73be84bb8eab8b63364633c07967b1fe1b754baa919fcbe892" Oct 04 06:09:35 crc kubenswrapper[4770]: I1004 06:09:35.239387 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:35 crc kubenswrapper[4770]: I1004 06:09:35.240498 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:36 crc kubenswrapper[4770]: I1004 06:09:36.287656 4770 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9p5z7" podUID="c14839de-6eee-42e9-a457-e66027e2d998" containerName="registry-server" probeResult="failure" output=< Oct 04 06:09:36 crc kubenswrapper[4770]: timeout: failed to connect service ":50051" within 1s Oct 04 06:09:36 crc kubenswrapper[4770]: > Oct 04 06:09:38 crc kubenswrapper[4770]: I1004 06:09:38.674392 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:09:38 crc kubenswrapper[4770]: E1004 06:09:38.675116 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:09:45 crc kubenswrapper[4770]: I1004 06:09:45.297634 4770 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:45 crc kubenswrapper[4770]: I1004 06:09:45.354278 4770 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:45 crc kubenswrapper[4770]: I1004 06:09:45.538285 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9p5z7"] Oct 04 06:09:47 crc kubenswrapper[4770]: I1004 06:09:47.292082 4770 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9p5z7" podUID="c14839de-6eee-42e9-a457-e66027e2d998" containerName="registry-server" containerID="cri-o://f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c" gracePeriod=2 Oct 04 06:09:47 crc kubenswrapper[4770]: I1004 06:09:47.781391 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:47 crc kubenswrapper[4770]: I1004 06:09:47.937783 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-utilities\") pod \"c14839de-6eee-42e9-a457-e66027e2d998\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " Oct 04 06:09:47 crc kubenswrapper[4770]: I1004 06:09:47.937917 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtvds\" (UniqueName: \"kubernetes.io/projected/c14839de-6eee-42e9-a457-e66027e2d998-kube-api-access-qtvds\") pod \"c14839de-6eee-42e9-a457-e66027e2d998\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " Oct 04 06:09:47 crc kubenswrapper[4770]: I1004 06:09:47.938039 4770 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-catalog-content\") pod \"c14839de-6eee-42e9-a457-e66027e2d998\" (UID: \"c14839de-6eee-42e9-a457-e66027e2d998\") " Oct 04 06:09:47 crc kubenswrapper[4770]: I1004 06:09:47.938701 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-utilities" (OuterVolumeSpecName: "utilities") pod "c14839de-6eee-42e9-a457-e66027e2d998" (UID: "c14839de-6eee-42e9-a457-e66027e2d998"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:09:47 crc kubenswrapper[4770]: I1004 06:09:47.943719 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14839de-6eee-42e9-a457-e66027e2d998-kube-api-access-qtvds" (OuterVolumeSpecName: "kube-api-access-qtvds") pod "c14839de-6eee-42e9-a457-e66027e2d998" (UID: "c14839de-6eee-42e9-a457-e66027e2d998"). InnerVolumeSpecName "kube-api-access-qtvds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.018520 4770 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c14839de-6eee-42e9-a457-e66027e2d998" (UID: "c14839de-6eee-42e9-a457-e66027e2d998"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.040310 4770 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtvds\" (UniqueName: \"kubernetes.io/projected/c14839de-6eee-42e9-a457-e66027e2d998-kube-api-access-qtvds\") on node \"crc\" DevicePath \"\"" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.040348 4770 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.040363 4770 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c14839de-6eee-42e9-a457-e66027e2d998-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.303312 4770 generic.go:334] "Generic (PLEG): container finished" podID="c14839de-6eee-42e9-a457-e66027e2d998" containerID="f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c" exitCode=0 Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.303367 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p5z7" event={"ID":"c14839de-6eee-42e9-a457-e66027e2d998","Type":"ContainerDied","Data":"f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c"} Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.303411 4770 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9p5z7" event={"ID":"c14839de-6eee-42e9-a457-e66027e2d998","Type":"ContainerDied","Data":"36fdb0f37a3faf5794384408e7807181c078323b9958e6347860375d9a0fd637"} Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.303432 4770 scope.go:117] "RemoveContainer" containerID="f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.303376 4770 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9p5z7" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.337509 4770 scope.go:117] "RemoveContainer" containerID="7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.338276 4770 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9p5z7"] Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.347653 4770 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9p5z7"] Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.366493 4770 scope.go:117] "RemoveContainer" containerID="76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.412463 4770 scope.go:117] "RemoveContainer" containerID="f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c" Oct 04 06:09:48 crc kubenswrapper[4770]: E1004 06:09:48.412947 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c\": container with ID starting with f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c not found: ID does not exist" containerID="f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.413095 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c"} err="failed to get container status \"f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c\": rpc error: code = NotFound desc = could not find container \"f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c\": container with ID starting with f1fc8d16aff439c2b652258f83e195d806fefde06f0b0e616d7e9373ac3d099c not found: ID does not exist" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.413192 4770 scope.go:117] "RemoveContainer" containerID="7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363" Oct 04 06:09:48 crc kubenswrapper[4770]: E1004 06:09:48.413558 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363\": container with ID starting with 7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363 not found: ID does not exist" containerID="7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.413577 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363"} err="failed to get container status \"7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363\": rpc error: code = NotFound desc = could not find container \"7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363\": container with ID starting with 7c847290d4316adda52af03f5bc8abfe4cd5285b5eed3fb7bab07639c0f85363 not found: ID does not exist" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.413591 4770 scope.go:117] "RemoveContainer" containerID="76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34" Oct 04 06:09:48 crc kubenswrapper[4770]: E1004 06:09:48.413863 4770 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34\": container with ID starting with 76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34 not found: ID does not exist" containerID="76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34" Oct 04 06:09:48 crc kubenswrapper[4770]: I1004 06:09:48.413957 4770 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34"} err="failed to get container status \"76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34\": rpc error: code = NotFound desc = could not find container \"76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34\": container with ID starting with 76a69eb48fa46d5fd4db9e069fd66766c3c4542dbe09c81e1941e2544a574a34 not found: ID does not exist" Oct 04 06:09:49 crc kubenswrapper[4770]: I1004 06:09:49.692081 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:09:49 crc kubenswrapper[4770]: E1004 06:09:49.692400 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:09:49 crc kubenswrapper[4770]: I1004 06:09:49.698300 4770 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c14839de-6eee-42e9-a457-e66027e2d998" path="/var/lib/kubelet/pods/c14839de-6eee-42e9-a457-e66027e2d998/volumes" Oct 04 06:10:03 crc kubenswrapper[4770]: I1004 06:10:03.674472 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:10:03 crc kubenswrapper[4770]: E1004 06:10:03.675434 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:10:18 crc kubenswrapper[4770]: I1004 06:10:18.674216 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:10:18 crc kubenswrapper[4770]: E1004 06:10:18.675123 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" Oct 04 06:10:31 crc kubenswrapper[4770]: I1004 06:10:31.675791 4770 scope.go:117] "RemoveContainer" containerID="ecd28ec902128ff1cb6930ea07de822e89d4f95f8bf6c95ae7e17888467ba588" Oct 04 06:10:31 crc kubenswrapper[4770]: E1004 06:10:31.676746 4770 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-h6msx_openshift-machine-config-operator(6bd3be93-1791-4cd1-b3ae-b4032548e93a)\"" pod="openshift-machine-config-operator/machine-config-daemon-h6msx" podUID="6bd3be93-1791-4cd1-b3ae-b4032548e93a" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070135150024442 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070135151017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070106622016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070106622015454 5ustar corecore